

i don’t think people in this forum would disagree with this move in 2018, as much as sentiments have changed. if you remove the political context and market moves from the equation, it is truly fascinating how these models work. GPT 2 was a crazy leap forward for language modeling, and the idea that a language model would be threatening middle class jobs wasn’t even on the table at that point. the idea that a pile of floating point numbers could write a React app is incredible, if politically fraught.
also, it wasn’t clear back then what OpenAI would become. they were a non-profit, and as clear as our hindsight is today this was before ChatGPT or any customer facing products were coming out of OpenAI.
i can’t be the only nerd in the room that has been fascinated by AI since i was a child only to face a reality where it’s not what i imagined it would be.





it’s kind of frustrating to have to keep explaining to people how these models work, mostly because of how intensely oversold they are.
on the one hand you have people who think it’s literally just a normal computer program doing database lookups with conditional logic and decision trees plus some sort of hand wavy magic. it’s not.
on the other hand you have people who think it’s a literal brain that can stub its toe and change the way it walks thereafter. it won’t.
every attempt at “agent memory” or whatever has thus far been desperate bullshit. i don’t care how many markdown files and vector databases and prompt engineering hacks you implement; you’ll never change the fact that these models have limited context and frozen weights. reading a markdown file or querying a database is not “remembering”.