John David Pressman's Tweets - August 2025

Back to Archive Index

πŸ”— John David Pressman 2025-08-02 16:30 UTC

Fortune teller who wants to get a good look at your hands because she's checking your digit ratio.

Likes: 8 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:00 UTC

@CharlesCMann @HSB_Lab LLMs are trained on the huge corpus of data, but what this actually means in practice is that each datapoint nudges its ability to "predict English text in general", it doesn't usually store that exact thing in its weights. So most details are forgotten during training.

Likes: 7 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:01 UTC

@CharlesCMann @HSB_Lab This can be mitigated by giving the model a web search capability so it can look up specific details from the web and incorporate those into its answer, but the answers are generated from a general prior over English that only remembers some details.

Likes: 7 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:02 UTC

@CharlesCMann @HSB_Lab It's kind of like how if I asked you exactly what you were doing at this time three weeks ago you probably couldn't tell me, but if I asked you about your last three weeks you could give me a sketch or overview of what you were doing. Your brain throws most daily details out.

Likes: 5 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:29 UTC

@CharlesCMann @GuiveAssadi @HSB_Lab Right so the thing that the training teaches is the ability to read those sources and parse them for meaning, and write a story about the details if it can retrieve them. But the details themselves are just too specific and niche to go into the general English text prior.

Likes: 2 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:29 UTC

@CharlesCMann @GuiveAssadi @HSB_Lab Does that make sense at all?

Likes: 0 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:34 UTC

@CharlesCMann @GuiveAssadi @HSB_Lab Yeah that part is basically the post-training recipe. How this works is you first teach it to predict text in general which produces a "base model", and using that model is kinda weird because it answers a question by trying to e.g. predict more of the question.

Likes: 2 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:35 UTC

@CharlesCMann @GuiveAssadi @HSB_Lab So you take that model and train it to answer questions with a special finetuning data set that is just users asking questions and getting answers so that it gives you answers instead of more question or going into a random forum thread or something.

Likes: 1 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:35 UTC

@CharlesCMann @GuiveAssadi @HSB_Lab And this special dataset is usually made by having contractors pretend to be the robot. So during OpenAssistant which was a community run system like this I saw a lot of people answer hard questions with "As an AI I can't do that" or "I don't know" but in character as the AI.

Likes: 1 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:36 UTC

@CharlesCMann @GuiveAssadi @HSB_Lab That's basically how it ends up like this. It learns from the human contractors to be lazy/confused about certain questions or to confidently state that nobody knows the answer because that's the style of how contractors would whiff on tasks that are too hard.

Likes: 1 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:42 UTC

@CharlesCMann @GuiveAssadi @HSB_Lab My explanation of a base model "predicting more of the question" probably wasn't very good so here's Kimi K2 with the instruction format ignored attempting to answer your question. As you can see this would be a pretty disorienting model for most users to use. https://t.co/e0DvOZG5Nf

Likes: 1 | Retweets: 0
πŸ”— John David Pressman 2025-08-04 21:47 UTC

@JimDMiller Why? GPT-5 probably won't do it for you.

Likes: 5 | Retweets: 0
πŸ”— John David Pressman 2025-08-05 00:23 UTC

PSA: If you really liked the style that CLIP Guided Diffusion and other old image models output in and are sad you can't get it on newer models, it's actually a look created by the use of gouache watercolor and airbrushing and you can prompt for it on newer models. https://t.co/cVfzcM9jY1

Likes: 33 | Retweets: 2
πŸ”— John David Pressman 2025-08-05 00:39 UTC

If you remove the default slop from the context window and add the term "airbrush": https://t.co/bgYLpXSR0W

Likes: 7 | Retweets: 0
πŸ”— John David Pressman 2025-08-05 08:23 UTC

This radicalized me when I read the earlier draft(?) a few years ago, but honestly not enough. I think maybe I need to print it out and put multiple copies of this part on the walls to help motivate me. x.com/lefthanddraft/…

Likes: 35 | Retweets: 0
πŸ”— John David Pressman 2025-08-05 08:31 UTC

@dogecahedron In practice? No.

Likes: 4 | Retweets: 0
πŸ”— John David Pressman 2025-08-05 08:56 UTC

@dogecahedron The cohort of "PauseAI" people who were present at the time to do so absolutely wanted to destroy GPT-4 and tried to ban any open GPT-4 type model as a stepping stone to that. In terms of policy that might actually happen these are the same movement.
1a3orn.com/sub/machine-le…

Likes: 6 | Retweets: 0
πŸ”— John David Pressman 2025-08-05 08:57 UTC

@dogecahedron "I want AGI banned FOREVER" is basically flavor text on top of "I want an AGI moratorium", in that we have a ban on human cloning that is indefinite and this will last exactly as long as people want it to and no longer, possibly just until someone breaks the taboo.

Likes: 4 | Retweets: 0
πŸ”— John David Pressman 2025-08-05 18:55 UTC

@JimDMiller Since you like prediction markets, how about you do this for me buddy:

Mark down that I told you GPT-5 probably won't obviate your need to write your thing, whatever it is, and then check if I was right after GPT-5 comes out in however many days. Or better yet a few weeks after.

Likes: 2 | Retweets: 0
πŸ”— John David Pressman 2025-08-06 20:17 UTC

@Xenoimpulse You don't get it, they just have to purity police harder and it'll work this time, pinkie promise.

Likes: 4 | Retweets: 0
πŸ”— John David Pressman 2025-08-13 02:00 UTC

Mu. x.com/tszzl/status/1… https://t.co/9G4TQyMTyH

Likes: 16 | Retweets: 0
πŸ”— John David Pressman 2025-08-13 02:00 UTC

Then again I forgot about the pro legacy switcher.
x.com/tszzl/status/1…

Likes: 9 | Retweets: 0

Want your own Twitter archive? Modify this script.

Twitter Archive by John David Pressman is marked with CC0 1.0