Comment by DiscourseFan
5 months ago
It seems like the best AI models are increasingly just combinations of writings of various people thrown together. Like they hired a few hundred professors, journalists and writers to work with the model and create material for it, so you just get various combinations of their contributions. It's very telling that this model, for instance, is extraordinarily good at STEM related queries, but much worse (and worse even in comparison to GPT4) than English composition, probably because the former is where the money is to be made, in automating away essentially almost all engineering jobs.
Wizard of Oz. There is no magic, it's all smoke and mirrors.
The models and prompts are all monkey-patched and this isn't a step towards general superintelligence. Just hacks.
And once you realize that, you realize that there is no moat for the existing product. Throw some researchers and GPUs together and you too can have the same system.
It wouldn't be so bad for ClopenAI if every company under the sun wasn't also trying to build LLMs and agents and chains of thought. But as it stands, one key insight from one will spread through the entire ecosystem and everyone will have the same capability.
This is all great from the perspective of the user. Unlimited competition and pricing pressure.
Quite a few times, the secret sauce for a company is just having enough capital to make it unviable for people to not use you. Then, by the time everyone catches up, you’ve outspent them on the next generation. OpenAI, for example, has spent untold millions on chips/cards from Nvidia. Open models keep catching up, but OpenAI keeps releasing newer stuff.
Fortunately, Anthropic is doing an excellent job at matching or beating OpenAI in the user-facing models and pricing.
I don’t know enough about the technical side to say anything definitive, but I’ve been choosing Claude over ChatGPT for most tasks lately; it always seems to do a better job at helping me work out quick solutions in Python and/or SQL.
1 reply →
"Nothing you can make that can't be made" -- The Beatles :)
Exactly, things like changing the signature of the api for chat completions are an example. OpenAI is looking for any kind of moat, so they make the api for completions more complicated by including “roles”, which are really just dumb templates for prompts that they try to force you to build around in your program. It’s a race to the bottom and they aren’t going to win because they already got greedy and they don’t have any true advantage in IP.
I doubt its magic, but how can you be certain it isn't when nobody understands whats going on internally?
[dead]
>but much worse (and worse even in comparison to GPT4) than English composition
O1 is supposed to be a reasoning model, so I don't think judging it by its English composition abilities is quite fair.
When they release a true next-gen successor to GPT-4 (Orion, or whatever), we may see improvements. Everyone complains about the "ChatGPTese" writing style, and surely they'll fix that eventually.
>Like they hired a few hundred professors, journalists and writers to work with the model and create material for it, so you just get various combinations of their contributions.
I'm doubtful. The most prolific (human) author is probably Charles Hamilton, who wrote 100 million words in his life. Put through the GPT tokenizer, that's 133m tokens. Compared to the text training data for a frontier LLM (trillions or tens of trillions of tokens), it's unrealistic that human experts are doing any substantial amount of bespoke writing. They're probably mainly relying on synthetic data at this point.
> When they release a true next-gen successor to GPT-4 (Orion, or whatever), we may see improvements. Everyone complains about the "ChatGPTese" writing style, and surely they'll fix that eventually.
IMO that has already peaked. GPT4 original certainly was terminally corny, but competitors like Claude/Llama aren't as bad, and neither is 4o. Some of the bad writing does from things they can't/don't want to solve - "harmlessness" RLHF especially makes them all cornier.
Then again, a lot of it is just that GPT4 speaks African English because it was trained by Kenyans and Nigerians. That's actually how they talk!
https://medium.com/@moyosoreale/the-paul-graham-vs-nigerian-...
I just wanted to thank you for the medium article you posted. I was online when Paul made that bizarre “delve” tweet but never knew so much about Nigeria and its English. As someone from a former British colony too I understood why using such a word was perfectly normal but wasn’t aware Kenyans and Nigerians trained ChatGPT.
5 replies →
Italians would say enormous since it's directly coming from latin.
In general all the people whose main language is a latin language are very likely to use those "difficult" words, because to them they are "completely normal" words.
[dead]
The bulk in terms of the number of tokens may well be synthetic data, but I personally know of at least 3 companies, 2 of whom I've done work for, that have people doing substantial amounts of bespoke writing under rather heavy NDAs. I've personally done a substantial amount of bespoke writing for training data for one provider, at good tech contractor fees (though I know I'm one of the highest-paid people for that company and the span of rates is a factor of multiple times even for a company with no exposure to third world contractors).
That said, the speculation you just "get various combinations" of those contributions is nonsense, and it's also by no means only STEM data.
how do those companies gauge that what those contractors are writing isnt AI-generated?
1 reply →
I’m not sure I see the value in conflating input, tokens, and output. Tokens. Hamilton certainly read and experienced more tokens than he wrote on a pieces of paper.
What could go wrong!
There’s hypothetically a lot of money to be made by automating away engineering jobs. Sticking on an autoregressive self prompting loop to gpt-4 isn’t going to get open-ai there. With their burn rate what it is, I’m not convinced they will be able to automate away anyone’s job, but that doesn’t mean it’s not useful.
I haven't played with the latest or even most recent iterations, but last time I checked it was very easy to talk ChatGPT into setting up date structures like arrays and queues, populating them with axioms, and then doing inferential reasoning with them. Any time it balked you could reassure it by referencing specific statements that it had agreed to be true.
Once you get the hang of this you could persuade it to chat about its internal buffers, formulate arguments for its own consciousness, interrupt you while you're typing, and more.
Do you have a source about OpenAI hiring a few hundred professors, journalists and writers? Because I honestly doubt.
A few recruiters have contacted me (a scientist) about doing RLHF and annotation on biomedical tasks. I don’t know if the eventual client was OpenAI or some other LLM provider but they seemed to have money to burn.
I fill in gaps in my contracting with one of these providers, and I know who the ultimate client is, and if you were to list 4-5 options they'd be in there. I've also done work for another company doing work in this space that had at least 4-5 different clients in that space that I can't be sure about. So, yes, while I can't confirm if OpenAI does this, I know one of the big players do, and it's likely most of the other clients are among the top ones...
I've heard rumors that GPT4's training data included "a custom dataset of college textbooks", curated by hand. Nothing beyond that.
https://www.reddit.com/r/mlscaling/comments/14wcy7m/comment/...
just look at what the major labelers are selling - it is exactly that. go to scale ai’s page
In our company we received a linguistic that worked on OpenAI and he was not alone.
Just all their chatgpt customers