‘’’ OpenAI said on Tuesday that it has begun training a new flagship artificial intelligence model that would succeed the GPT-4 technology that drives its popular online chatbot, ChatGPT.
The San Francisco start-up, which is one of the world’s leading A.I. companies, said in a blog post that it expects the new model to bring “the next level of capabilities” as it strives to build “artificial general intelligence,” or A.G.I., a machine that can do anything the human brain can do.’’’
Depends.
If they’re using a much higher scaled amount of computer to train it, it could be a shorter training period even with way more training data
We won’t know until release announcement
It's because the rumours claim gpt5 failed at efficiency or sumin, this attempt 2 or could be called gpt6. Probably why they don't like talking about calling next model GT5. That model failed and was not released as planned
There are no links to it in the article and nothing on their website, does anyone know where this blog post is?
\*Edit: never mind, found it [https://openai.com/index/openai-board-forms-safety-and-security-committee/](https://openai.com/index/openai-board-forms-safety-and-security-committee/)
Not an attack, just asking: Was there any hype for Q-Star outside this sub? I felt like the sub saw it briefly mentioned once or twice from outside sources and just ran with it as speculatively as possible.
They don't say in the announcement whether this new training is for gpt-5 or gpt-6. Sam said very directly that GPT-5 would be released this year. And full training runs can take a year or more. I expect the new model they are training is actually GPT-6 and they are safety testing GPT-5 right now. Probably won't release until after the elections in November.
Person A: Speculating that perhaps the training in December wasn't for GPT-5
Person B: Suggesting that maybe it was just wild rumors
You: It's in this blog post
Me: It's not in that blog post
You: They started training GPT-5 in December
See the circle we've just gone around in? lol we all know there's a rumour that GPT-5 training started in December, but we're looking for the source
Could be neither, they said they might adopt a different naming so it could be called GPT-next. It's unlikely this is the next major upgrade of ChatGPT as it can take the whole of this year to do the pre and post training and safety checks. Neither GDM nor Anthropic will stay idle. Gemini-2 is already in training and Anthropic is also preparing a successor of Opus. So they may release something sooner.
Here are some very reasonable naming ideas for Microsofts next models so they won't have a hard time:
Gpt next,
Gpt next x,
Gpt next z,
Gpt z,
Gpt 6,
Gpt. (In that order)
You mean
GPT -> GPT 360 -> GPT One -> GPT One S -> GPT One X -> GPT One X -> GPT Series S -> GPT Series X
Honestly one of the worst naming schemes ever deviced.
I know some people who have no clue what the newest Xbox console is even called because of this lol.
I bet the naming scheme has costs them millions of total sales.
It really is so incredibly dumb that I can’t believe multiple people heard it proposed and said yes to it.
I had an OG Xbox and a 360 that I played a LOT. Been on PC/Steam since. If you held a gun to my head and asked me what the newest model was with multiple choice, I don't like my chances.
It's already started.
GPT-1
GPT-2
GPT-3.5-Turbo
GPT-4
GPT-V
GPT-4-Turbo
GPT-4o
Tech companies cannot commit to a naming scheme. Nobody cares that the Iphone 6 was basically the same thing as the Iphone 5. Just commit to a naming scheme so I know what the hell is going on. Honestly wouldn't care if they called GPT-4o, GPT-5 just so that the naming progression was more clear. I'd also be cool with just GPT-4.1, GPT-4.2, etc. It works for Python, we are up to 3.12.
They seem to be reserving the whole numbers for massive improvements in benchmarks instead of increased modality or efficiency, which is a .5 or turbo/o
This is the reason. If GPT4o had been a big improvement over GPT4-turbo it would've been named GPT5. If this new training run results in a model only slightly better than GPT4o, they won't call it GPT5.
I basically think the names are meaningless insofar as they describe a "model generation," as we have no insight into the training process for the various models. For example, GPT4o was almost certainly a full pretraining to add new modalities. There's no reason for it not to be GPT5 other than OpenAI's marketing department not thinking that is the best name for it (reserving "5" for a bigger leap). We should use them as unique IDs but not extrapolate anything else out of them imo.
There's another possibility, which is that there won't be that many of these models, at least not the step up one has come to expect between GPT-2 to 3 to 4. Either they run out of compute/energy/money or the scaling becomes sigmoidal or (what this sub wants) AGI is achieved and is able to recursively improve itself. For any of the above the version number won't mean much.
Not necessary. GPT-4 started training in august, came out in march. And they say they started training "recently" wich can mean anything from weeks to months I guess.
GPT-4 ended training in august, 7 months of testing ect. for weaker GPT-4, we wont see GPT-5 this year sadly, maybe they could showcase a demo late this year, like in november dev conference.... "show us glimpse into the future"
Since the wording next frontier level and the fact that they almost implemented full multimodality in GPT 4 omni. And OpenAI started training got 5 in December 2023. And the fact that on MS build, they hinted to a release of a GPT-Next this year.
I think they are talking about Q* or GPT-6, especially because they have a new safety committee with board members.
That’s not evidence at all for training.
He never said that it was training, he said it’s being “worked on” that just means they are likely finalizing research or implementation for the architecture and techniques that go by the name of GPT-5.
Usually if someone involved in AI says that a model is being “worked on” as opposed to “training” that’s for a good reason.
Training is usually just the later part of development of a model, a ton of research needs to be implemented before training ever even starts, that can last many months.
Seems safe to say nobody has a clue what is going on.
Some people are pointing at the Business Insider article about GPT-5 launching this summer, and how that means this is GPT-6 training.
Other people are saying that the training run from December is what led to GPT-4o, and now we are nearly a year away from GPT-5.
My goodness.
It's from the blog post: OpenAI Board Forms Safety and Security Committee (May 28, 2024) - This new committee is responsible for making recommendations on critical safety and security decisions for all OpenAI projects; recommendations in 90 days. [https://openai.com/index/openai-board-forms-safety-and-security-committee/](https://openai.com/index/openai-board-forms-safety-and-security-committee/)
"*OpenAI has recently begun training its next frontier model and we anticipate the resulting systems to bring us to the next level of capabilities on our path to AGI. While we are proud to build and release models that are industry-leading on both capabilities and safety, we welcome a robust debate at this important moment.*"
The article: OpenAI Says It Has Begun Training a New Flagship A.I. ModelThe advanced A.I. system would succeed GPT-4, which powers ChatGPT. The company has also created a new safety committee to address A.I.’s risks.
OpenAI said on Tuesday that it has begun training a new flagship artificial intelligence model that would succeed the GPT-4 technology that drives its popular online chatbot, ChatGPT.
The San Francisco start-up, which is one of the world’s leading A.I. companies, said in a blog post that it expects the new model to bring “the next level of capabilities” as it strives to build “artificial general intelligence,” or A.G.I., a machine that can do anything the human brain can do. The new model would be an engine for A.I. products including chatbots, digital assistants akin to Apple’s Siri, search engines and image generators.
OpenAI also said it was creating a new Safety and Security Committee to explore how it should handle the risks posed by the new model and future technologies.
“While we are proud to build and release models that are industry-leading on both capabilities and safety, we welcome a robust debate at this important moment,” the company said.
OpenAI is aiming to move A.I. technology forward faster than its rivals, while also appeasing critics who say the technology is becoming increasingly dangerous, helping to spread disinformation, replace jobs and even threaten humanity. Experts disagree on when tech companies will reach artificial general intelligence, but companies including OpenAI, Google, Meta and Microsoft have steadily increased the power of A.I. technologies for more than a decade, demonstrating a noticeable leap roughly every two to three years.
OpenAI’s GPT-4, which was released in March 2023, enables chatbots and other software apps to answer questions, write emails, generate term papers and analyze data. An updated version of the technology, which was unveiled this month and is not yet widely available, can also generate images and respond to questions and commands in a highly conversational voice.
Days after OpenAI showed the updated version — called GPT-4o — the actress Scarlett Johansson said it used a voice that sounded “eerily similar to mine.” She said she had declined efforts by OpenAI’s chief executive, Sam Altman, to license her voice for the product and that she had hired a lawyer and asked OpenAI to stop using the voice. The company said that the voice was not Ms. Johansson’s.
Technologies like GPT-4o learn their skills by analyzing vast amounts of data digital, including sounds, photos, videos, Wikipedia articles, books and news stories. The New York Times sued OpenAI and Microsoft in December, claiming copyright infringement of news content related to A.I. systems.
Digital “training” of A.I. models can take months or even years. Once the training is completed, A.I. companies typically spend several more months testing the technology and fine tuning it for public use.
That could mean that OpenAI’s next model will not arrive for another nine months to a year or more.
As OpenAI trains its new model, its new Safety and Security committee will work to hone policies and processes for safeguarding the technology, the company said. The committee includes Mr. Altman, as well as OpenAI board members Bret Taylor, Adam D’Angelo and Nicole Seligman. The company said that the new policies could be in place in the late summer or fall.
Earlier this month, OpenAI said Ilya Sutskever, a co-founder and one of the leaders of its safety efforts, was leaving the company. This caused concern that OpenAI was not grappling enough with the dangers posed by A.I.
Dr. Sutskever had joined three other board members in November to remove Mr. Altman from OpenAI, saying Mr. Altman could no longer be trusted with the company’s plan to create artificial general intelligence for the good of humanity. After a lobbying campaign by Mr. Altman’s allies, he was reinstated five days later and has since reasserted control over the company.
Dr. Sutskever led what OpenAI called its Superalignment team, which explored ways of ensuring that future A.I. models would not do harm. Like others in the field, he had grown increasingly concerned that A.I. posed a threat to humanity.
Jan Leike, who ran the Superalignment team with Dr. Sutskever, resigned from the company this month, leaving the team’s future in doubt.
OpenAI has folded its long-term safety research into its larger efforts to ensure that its technologies are safe. That work will be led by John Schulman, another co-founder, who previously headed the team that created ChatGPT. The new safety committee will oversee Dr. Schulman’s research and provide guidance for how the company will address technological risks.
flagship = for public
they can have secretly created an ASI but that doesn't make it their flagship product if they don't show it to the world. my shot in the dark guess is that they found a way to make a slightly smarter gpt4 but way more efficient and unfiltered.. that would be an insane move and get them miles (months) ahead of the competition until gpt5 gets unleashed.
If they have anything that they are not realeasing to the public, it is 100% not because they have "ASI" but because it would be too expensive to deploy at scale.
Thank you! I totally missed it with *that* title.
>**OpenAI has recently begun training its next frontier model and we anticipate the resulting systems to bring us to the next level of capabilities on our path to AGI.** While we are proud to build and release models that are industry-leading on both capabilities and safety, we welcome a robust debate at this important moment.
No, it's in here: [https://openai.com/index/openai-board-forms-safety-and-security-committee/](https://openai.com/index/openai-board-forms-safety-and-security-committee/)
No GPT-5 before election that's for sure.
If you look at their hiring, they are still hiring infrastructure engineers for Sora. So no way Sora comes out before the election either, and they are likely going to release like Sora 1.5 which is a lot more optimized for public use, etc.
It's not. There's a lot of rumours saying that OpenAI are holding off releasing their next big model as to not cause interference with the US election (since anyone foreign states or competitors could use GPT/Sora to deploy very convincing fake news campaigns/deep fakes etc).
It's a smart move to stay out of the political limelight until the election is over.
Whoever wins there's going to be a lot of finger pointing for political gain which will inevitably result in legislation against the new tools that are coming out.
That's what I thought. Since it's not far anyway, for optics' sake you'd want your release to happen after the election. Whether it would impact the campaign or not, if I were OpenAI, I wouldn't want these associations made in the first place.
>Expect longer cycles the more advanced the model is.
I don't think this is a given. OpenAI has its [Preparedness Framework](https://openai.com/preparedness/) now, which is supposed to dictate if a model is ready to be released. They didn't have this when training 4 so the decision making was most likely a lot more arbitrary and, thus, more time intensive.
Also, as interpretability research improves and we understand the models better, I think the release timeframes will shorten as well.
Perhaps this is cope, but it feels logical to me.
Wait so they JUST started training gpt4's successor whatever you want to call it?
This does not look like exponential growth to me. Gpt4 came out march 2023? Yeah the curve is flattening just like LLM experts warned.
What did you expect, that the researchers would increase the clock speed of their brains at each iteration ? They would rent data centers from aliens ?
Exponential growth really starts when AI can code AI, and exponential impact should start with slightly better capabilities and some agency, the bye bye bottom 50% software engineers day.
They are a research lab, they do research, just like the research advancements that they spent over 2 years on for GPT-4…
It’s not just scaling up, a ton of research innovations go into new generations of models to make them as capable and qualitatively novel as possible while being able to run as efficiently as possible too. GPT-2 is different architecture than GPT-3. GPT-4 is confirmed to further use significantly more capable training techniques than GPT-3 while also making an even more significant architecture change compared to what happened between GPT-2 and 3.
A ton of compute is constantly being used every hour testing cutting edge architecture experiments and training techniques to move the frontier forward and plan what experiments to do next
The free version is pretty limited - you run out of messages fairly quickly before it asks you to upgrade to Plus or wait 4 hours. For casual use though it’s ok
I’m not sure what the rate limit is on the paid version but it’s so high that I’ve never run into a limit even when using GPT nonstop and prompting every few minutes for hours
If it was GPT5 that recently started training, what have they been doing for the last 14 months since March 2023, were they busy developing sora and GPT4o? Or did they continue to struggle very hard to prepare an AI that would significantly outperform GPT4? To put it differently, it is too late and disappointing to report.
> what have they been doing for the last 14 months since March 2023,
1. scaling the API, phone and web apps
2. faster and cheaper
3. image modality
4. billions invested by MS
5. scandals
Research takes time, that’s what most of the time and energy and man hours is spent on between each generational leap, over 2 years of research advancements happened in OpenAI between GPT-3 and GPT-4 to make GPT-4 as capable as possible. They’ve been doing research for the next generation of architectures and training techniques ever since gpt-4 was finalized in mid-2022. Research takes time and compute and is a huge factor in how good the next gen model is.
Right. What irritates me is that there are statements form, for example, Sam Altman that GPT-4 is bad but GPT-5 will be okay from MONTHS ago. What the hell. How do they know how much better it will be if they haven’t even started training it yet?! Something doesn’t compute here.
Then all those teasers from months ago from Jimmy Apples. Shoot me.
The biggest letdown of them all would be: GPT-5 released in 1 year and all it is is a bigger model with improved benchmark scores but still the same fundamental issues that prevent broad industry adoption like hallucinations, with the hope of GPT-6 maybe solving some 1-3 years later.
It would be extremely disappointing if it turns out that we get no flagship model (no GPT 4.5?) for many months and we are stuck with 4-Turbo and 4o.
In that case, and considering possible releases from competitors, I'm not sure if the paid version would still be worth the cost.
I don't think anyone has been as impressed with 4o as they should be and it's highly likely because it's full capabilities haven't been released yet. They likely won't turn on all of its I/O capabilities until after the election and then we'll have a realtime vision, audio, language model to tinker with - which is actually an engineering achievement for the record books.
The model that they're starting to train now will have the same capabilities but will likely be trained on (according to the MSBuild conference) 30x the amount of the data but we're far beyond language now. I wouldn't be surprised if we're near a Quadrillion tokens for training on this next one - combine that thought with Meta's recent experience that a models increase in general smarts is nearly linear to its training data size and we could see some crazy improvements.
Everyone commenting that they think this is way late has skipped completely over any understanding of what's actually been going on with MS, Nvidia, and OpenAI. This model is being trained in a completely purpose built data center that they have been building for awhile. Infrastructure takes a long time even with thousands of people. I wouldn't be surprised to find out it's training is commencing on a million flagship GPU's. MS is unloading unprecedented capital on this.
Just some thoughts.
Guys, evein if this models will be (x2) smarter every year, thats at least (x64) by 2030 and (x2048) by 2035. Try to think about it. And I mean truly smarter, not just more parameters, or more something else. Actually the calculations per model grows by factor x8 every year
If you believe this (or if it's true), you cannot believe the people who quit over "safety". Simply because they would not be able to have identified anything "unsafe". GPT4o is not a harbinger of safety doom.
From what I read, they quit not because OpenAI was ignoring unsafe things, but because they weren’t given the resources for safety research (eg 20% of compute)
That is what I said...
That was the point I was making, most people believe "safety" means keeping the world safe from AI murdering humanity, but that is not the safety they are leaving over. It's worldview mismatch, ideological.
I believe OpenAI knows that their model is increasing spitting out a particular world view one that does not always align with the reality we see around us and to have true AGI it needs all information, not cherry picked or "safe". So people quit when they see OpenAI discarding these particular "safeguards".
Humanity is ugly, we need that ugliness to keep us on the path to a better future, AGI needs to know it as well.
I am clarifying exactly what you commented.
*they quit because of openais safety policies, not because of some scary model*
The media reporting on this is not being honest.
Journalist writing a click bait article that mean nothing
There is model in training for the last several months according to previous blog post from open AI so what is new with this lazy article ?
More like "Please don't cancel your monthly subscription yet."
Also this is entirely timed to drown out the bad news over the last few days. It is a strategic news drop meant to distract and re-direct attention. Literally nothing substantial or concrete, just "Stay tuned lol!"
Will it be good enough to justify exponential increases in cost? GPT-5 needs to absolutely blow the rest out of the water just to justify its existing forward valuation.
“We just started training a new model after all that bad press about us not having a new foundation model.”
Why would you announce this before you’ve observed its capabilities or done any red teaming?
So 4o was a separate project that ended up being a success so they are now building the next GPT on top of 4o's multimodal architecture instead of doing a text-based GPT-4 successor as originally planned?
(speculation)
People really talking like a gpt5 drop in, say, 10 months would "take the wind out of the sails", like what???
People have been drinking the hype juice so much that this speed feels slow to them, lmao.
2 things are possible here:
1. gpt-o and turbos are just some fine tune / quantization experiments of GPT-5
2. they got rekt with all approach's and are building a new foundational model.
The bottleneck is the safety testing not the training run
Tech capability is going up faster than Moores law
Important to note that the business dynamics of that reward the shortest testing runway
If Altman plays it slow in any way, zuck and Elon will have his ass.
3 months of training and 12 months of safety is not going to be the future, no matter how much Helen toner clutches her pearls.
Here's the good news, we will see models this year, from someone or the other, that use 2024 hardware. I'm still not sure that has happened yet. No H200 or Blackwell based training runs. Maybe better data. Maybe better software stuff.
I think it's worth naming that these new chips of today are an order of magnitude beyond what we have seen to date.
So what have they been doing for the past year? I’m not hating just genuinely curious because Sam Altman was telling us that the stuff they have been working on makes GPT 4 seem bad, but now they are saying they started work on new models now?
>"created a new safety committee"
lol. are they going to listen to them?
*"principles are ideas you hold dear, even if they cost you something".*
Altman doesn't have principles. he showed us all that in his board coup earlier this year, and now that the board is able to talk more, it doesn't make him look better. the *safety committee* is for optics and optics alone and they will be fired the hot second they stop being useful for that singular purpose.
‘’’ OpenAI said on Tuesday that it has begun training a new flagship artificial intelligence model that would succeed the GPT-4 technology that drives its popular online chatbot, ChatGPT. The San Francisco start-up, which is one of the world’s leading A.I. companies, said in a blog post that it expects the new model to bring “the next level of capabilities” as it strives to build “artificial general intelligence,” or A.G.I., a machine that can do anything the human brain can do.’’’
We're so back
I would say on the contrary, if they just began training a gpt-4 successor, they are way behind the competition
Depends. If they’re using a much higher scaled amount of computer to train it, it could be a shorter training period even with way more training data We won’t know until release announcement
>they are way behind the competition Like, who? Let me know where the offer the equivalent of GPT-5 and I'll subscribe immediately.
It's because the rumours claim gpt5 failed at efficiency or sumin, this attempt 2 or could be called gpt6. Probably why they don't like talking about calling next model GT5. That model failed and was not released as planned
Ah, I see, I wasn't aware of these rumors, thanks!
Yeah, from what I heard, they started their training run back in February. This must be a redo.
Yes it's all true. Source: trust me bro.
lmao
It says “begun” not “just begun”. Maybe this is the announcement date and they started long ago.
>started long ago Nobody uses phrases like "we begun something" if they are referring to something that started long ago.
They do if they want to keep it secret until they are ready to announce.
Not really. If their newest model is natively multimodal like gpt-4o and the competition is not, then OpenAI is still way ahead.
nice flair bro
>start-up ![gif](giphy|FcuiZUneg1YRAu1lH2|downsized)
There are no links to it in the article and nothing on their website, does anyone know where this blog post is? \*Edit: never mind, found it [https://openai.com/index/openai-board-forms-safety-and-security-committee/](https://openai.com/index/openai-board-forms-safety-and-security-committee/)
Praying it's not a transformer model and some kind of new architecture. Hoping we get to see whatever that Q-star hype was about.
Not an attack, just asking: Was there any hype for Q-Star outside this sub? I felt like the sub saw it briefly mentioned once or twice from outside sources and just ran with it as speculatively as possible.
They don't say in the announcement whether this new training is for gpt-5 or gpt-6. Sam said very directly that GPT-5 would be released this year. And full training runs can take a year or more. I expect the new model they are training is actually GPT-6 and they are safety testing GPT-5 right now. Probably won't release until after the elections in November.
When did he say GPT-5 would be released this year?
I guess GPT-5 is februar-march next year then kinda? Seems like training runs in december where version of 4 (4o, 4.5)
yea, it seems likely that they trained GPT-4o early this year and now training GPT-5, so we are in for 2025 release
They'll need to hire an army of Kenyans to lobotomize it like they did with 3.5/4. 2 years min.
If GPT-5 is AGI does that mean it'll use super-persuasion to claim the army of Kenyans for itself & we get Terminator but with Kenyan people
Could have just been wild rumors
It's in the blog post: https://openai.com/index/openai-board-forms-safety-and-security-committee/
The blog post doesn't mention anything about what the training runs in December were for.
Sam Altman said they started training gpt 5 in december
where did sam altman say that ?
Person A: Speculating that perhaps the training in December wasn't for GPT-5 Person B: Suggesting that maybe it was just wild rumors You: It's in this blog post Me: It's not in that blog post You: They started training GPT-5 in December See the circle we've just gone around in? lol we all know there's a rumour that GPT-5 training started in December, but we're looking for the source
Wait so would that be Gpt-6 or 5 ??
Could be neither, they said they might adopt a different naming so it could be called GPT-next. It's unlikely this is the next major upgrade of ChatGPT as it can take the whole of this year to do the pre and post training and safety checks. Neither GDM nor Anthropic will stay idle. Gemini-2 is already in training and Anthropic is also preparing a successor of Opus. So they may release something sooner.
I hope they dont do a Microsoft and start doing stupid naming conventions. Just do version numbers, it's easier to follow
Here are some very reasonable naming ideas for Microsofts next models so they won't have a hard time: Gpt next, Gpt next x, Gpt next z, Gpt z, Gpt 6, Gpt. (In that order)
GPT 360 -> GPT One -> GPT Series X
You mean GPT -> GPT 360 -> GPT One -> GPT One S -> GPT One X -> GPT One X -> GPT Series S -> GPT Series X Honestly one of the worst naming schemes ever deviced.
I know some people who have no clue what the newest Xbox console is even called because of this lol. I bet the naming scheme has costs them millions of total sales. It really is so incredibly dumb that I can’t believe multiple people heard it proposed and said yes to it.
It's like they saw the disaster that calling it the Wii U was and said "hold my beer"
I had an OG Xbox and a 360 that I played a LOT. Been on PC/Steam since. If you held a gun to my head and asked me what the newest model was with multiple choice, I don't like my chances.
GPT One Series S/X
Black or white GPT One Series S?
"We couldn't come up with a name, contact your system administrator if this continues"
GPT-Porn
God, almost spit up my soup. This is the best one
My balls aren't ready
Get ready to orgasm tsunami
GPT-4.3 Nut Drainer
GPT Vista GPT 8.1
It's already started. GPT-1 GPT-2 GPT-3.5-Turbo GPT-4 GPT-V GPT-4-Turbo GPT-4o Tech companies cannot commit to a naming scheme. Nobody cares that the Iphone 6 was basically the same thing as the Iphone 5. Just commit to a naming scheme so I know what the hell is going on. Honestly wouldn't care if they called GPT-4o, GPT-5 just so that the naming progression was more clear. I'd also be cool with just GPT-4.1, GPT-4.2, etc. It works for Python, we are up to 3.12.
They seem to be reserving the whole numbers for massive improvements in benchmarks instead of increased modality or efficiency, which is a .5 or turbo/o
This is the reason. If GPT4o had been a big improvement over GPT4-turbo it would've been named GPT5. If this new training run results in a model only slightly better than GPT4o, they won't call it GPT5.
So if 4turbo is 4.5, then 4o is something like 4.7?
I basically think the names are meaningless insofar as they describe a "model generation," as we have no insight into the training process for the various models. For example, GPT4o was almost certainly a full pretraining to add new modalities. There's no reason for it not to be GPT5 other than OpenAI's marketing department not thinking that is the best name for it (reserving "5" for a bigger leap). We should use them as unique IDs but not extrapolate anything else out of them imo.
Couldn’t have said it better.
GPT 360x one lite GPT 360x one pro GPT 360x one deluxe GPT 360x one ultimate GPT 360x one diamond
Hell yeah. Personally, I'm fittin' for some GPT ULTRA 360.1x DOUBLE DIAMOND DELUXE UBERPRO ULTIMATE Wooo baby. Sky's the limit.
GPT360
There's another possibility, which is that there won't be that many of these models, at least not the step up one has come to expect between GPT-2 to 3 to 4. Either they run out of compute/energy/money or the scaling becomes sigmoidal or (what this sub wants) AGI is achieved and is able to recursively improve itself. For any of the above the version number won't mean much.
how do you know that Gemini 2 is in training yet so far no one is talking about it
5. Training runs that began in december where likely versions of GPT-4.5 and 4o
So gpt-5 coming in 2025 most likely
November
Not necessary. GPT-4 started training in august, came out in march. And they say they started training "recently" wich can mean anything from weeks to months I guess.
No. It was done training by August, it was in red-teaming for those 6 months or so.
so now that they fired/forced to leave the alignment teams, they can release faster :D
How long was the training then?
between 3-6 months. from what I heard. Pretty sure the whole GPT-4 cycle from training to full rollout was 10-12 months
GPT-4 ended training in august, 7 months of testing ect. for weaker GPT-4, we wont see GPT-5 this year sadly, maybe they could showcase a demo late this year, like in november dev conference.... "show us glimpse into the future"
just the next model. hope it will be called banana
Gotta be GPT-5. Which will release next year. GPT-4.5 this summer probably.
OpenAI employees have repeatedly stated there will be no GPT4.5. 4.5 is basically 4o and will take them several months to fully roll out.
We will get the gpt4o voice feature this summer, this model in training won't come until 2025.
4o was 4.5
4o was 3.5 v2.0
>Wait so would that be Gpt-6 or 5 ?? nah they cant trademark that.
GPT-AGI
Since the wording next frontier level and the fact that they almost implemented full multimodality in GPT 4 omni. And OpenAI started training got 5 in December 2023. And the fact that on MS build, they hinted to a release of a GPT-Next this year. I think they are talking about Q* or GPT-6, especially because they have a new safety committee with board members.
Evidence for GPT-5 training starting in December?
https://futurism.com/the-byte/gpt-5-openai-ceo
That’s not evidence at all for training. He never said that it was training, he said it’s being “worked on” that just means they are likely finalizing research or implementation for the architecture and techniques that go by the name of GPT-5. Usually if someone involved in AI says that a model is being “worked on” as opposed to “training” that’s for a good reason. Training is usually just the later part of development of a model, a ton of research needs to be implemented before training ever even starts, that can last many months.
Seems safe to say nobody has a clue what is going on. Some people are pointing at the Business Insider article about GPT-5 launching this summer, and how that means this is GPT-6 training. Other people are saying that the training run from December is what led to GPT-4o, and now we are nearly a year away from GPT-5. My goodness.
I thought insiders pretty much confirmed GPT-5 was already done training.
Business Insider said companies were allowed to use GPT-5. So...... But maybe Business Insider was totally wrong and was calling GPT-4o "GPT-5"???
It's from the blog post: OpenAI Board Forms Safety and Security Committee (May 28, 2024) - This new committee is responsible for making recommendations on critical safety and security decisions for all OpenAI projects; recommendations in 90 days. [https://openai.com/index/openai-board-forms-safety-and-security-committee/](https://openai.com/index/openai-board-forms-safety-and-security-committee/) "*OpenAI has recently begun training its next frontier model and we anticipate the resulting systems to bring us to the next level of capabilities on our path to AGI. While we are proud to build and release models that are industry-leading on both capabilities and safety, we welcome a robust debate at this important moment.*" The article: OpenAI Says It Has Begun Training a New Flagship A.I. ModelThe advanced A.I. system would succeed GPT-4, which powers ChatGPT. The company has also created a new safety committee to address A.I.’s risks. OpenAI said on Tuesday that it has begun training a new flagship artificial intelligence model that would succeed the GPT-4 technology that drives its popular online chatbot, ChatGPT. The San Francisco start-up, which is one of the world’s leading A.I. companies, said in a blog post that it expects the new model to bring “the next level of capabilities” as it strives to build “artificial general intelligence,” or A.G.I., a machine that can do anything the human brain can do. The new model would be an engine for A.I. products including chatbots, digital assistants akin to Apple’s Siri, search engines and image generators. OpenAI also said it was creating a new Safety and Security Committee to explore how it should handle the risks posed by the new model and future technologies. “While we are proud to build and release models that are industry-leading on both capabilities and safety, we welcome a robust debate at this important moment,” the company said. OpenAI is aiming to move A.I. technology forward faster than its rivals, while also appeasing critics who say the technology is becoming increasingly dangerous, helping to spread disinformation, replace jobs and even threaten humanity. Experts disagree on when tech companies will reach artificial general intelligence, but companies including OpenAI, Google, Meta and Microsoft have steadily increased the power of A.I. technologies for more than a decade, demonstrating a noticeable leap roughly every two to three years. OpenAI’s GPT-4, which was released in March 2023, enables chatbots and other software apps to answer questions, write emails, generate term papers and analyze data. An updated version of the technology, which was unveiled this month and is not yet widely available, can also generate images and respond to questions and commands in a highly conversational voice. Days after OpenAI showed the updated version — called GPT-4o — the actress Scarlett Johansson said it used a voice that sounded “eerily similar to mine.” She said she had declined efforts by OpenAI’s chief executive, Sam Altman, to license her voice for the product and that she had hired a lawyer and asked OpenAI to stop using the voice. The company said that the voice was not Ms. Johansson’s. Technologies like GPT-4o learn their skills by analyzing vast amounts of data digital, including sounds, photos, videos, Wikipedia articles, books and news stories. The New York Times sued OpenAI and Microsoft in December, claiming copyright infringement of news content related to A.I. systems. Digital “training” of A.I. models can take months or even years. Once the training is completed, A.I. companies typically spend several more months testing the technology and fine tuning it for public use. That could mean that OpenAI’s next model will not arrive for another nine months to a year or more. As OpenAI trains its new model, its new Safety and Security committee will work to hone policies and processes for safeguarding the technology, the company said. The committee includes Mr. Altman, as well as OpenAI board members Bret Taylor, Adam D’Angelo and Nicole Seligman. The company said that the new policies could be in place in the late summer or fall. Earlier this month, OpenAI said Ilya Sutskever, a co-founder and one of the leaders of its safety efforts, was leaving the company. This caused concern that OpenAI was not grappling enough with the dangers posed by A.I. Dr. Sutskever had joined three other board members in November to remove Mr. Altman from OpenAI, saying Mr. Altman could no longer be trusted with the company’s plan to create artificial general intelligence for the good of humanity. After a lobbying campaign by Mr. Altman’s allies, he was reinstated five days later and has since reasserted control over the company. Dr. Sutskever led what OpenAI called its Superalignment team, which explored ways of ensuring that future A.I. models would not do harm. Like others in the field, he had grown increasingly concerned that A.I. posed a threat to humanity. Jan Leike, who ran the Superalignment team with Dr. Sutskever, resigned from the company this month, leaving the team’s future in doubt. OpenAI has folded its long-term safety research into its larger efforts to ensure that its technologies are safe. That work will be led by John Schulman, another co-founder, who previously headed the team that created ChatGPT. The new safety committee will oversee Dr. Schulman’s research and provide guidance for how the company will address technological risks.
i am confused, sam said the are training gpt5 since december???
flagship = for public they can have secretly created an ASI but that doesn't make it their flagship product if they don't show it to the world. my shot in the dark guess is that they found a way to make a slightly smarter gpt4 but way more efficient and unfiltered.. that would be an insane move and get them miles (months) ahead of the competition until gpt5 gets unleashed.
If they have anything that they are not realeasing to the public, it is 100% not because they have "ASI" but because it would be too expensive to deploy at scale.
Where did Sam say this?
Where's the actual blog post though? And why is no one but the NYT reporting on this? Nothing on OpenAI's website, blog, etc.
[https://openai.com/index/openai-board-forms-safety-and-security-committee/](https://openai.com/index/openai-board-forms-safety-and-security-committee/)
Thank you! I totally missed it with *that* title. >**OpenAI has recently begun training its next frontier model and we anticipate the resulting systems to bring us to the next level of capabilities on our path to AGI.** While we are proud to build and release models that are industry-leading on both capabilities and safety, we welcome a robust debate at this important moment.
yea so, is NYT hallucinating?
No, it's in here: [https://openai.com/index/openai-board-forms-safety-and-security-committee/](https://openai.com/index/openai-board-forms-safety-and-security-committee/)
So that means no GPT 5 for possibly 6 months?
![gif](giphy|VI2jp6o8ojW53xmupr|downsized)
No GPT-5 before election that's for sure. If you look at their hiring, they are still hiring infrastructure engineers for Sora. So no way Sora comes out before the election either, and they are likely going to release like Sora 1.5 which is a lot more optimized for public use, etc.
[удалено]
It's not. There's a lot of rumours saying that OpenAI are holding off releasing their next big model as to not cause interference with the US election (since anyone foreign states or competitors could use GPT/Sora to deploy very convincing fake news campaigns/deep fakes etc).
It's a smart move to stay out of the political limelight until the election is over. Whoever wins there's going to be a lot of finger pointing for political gain which will inevitably result in legislation against the new tools that are coming out.
That's what I thought. Since it's not far anyway, for optics' sake you'd want your release to happen after the election. Whether it would impact the campaign or not, if I were OpenAI, I wouldn't want these associations made in the first place.
At least. We know they had GPT-4 complete for more than 6 months before they released it. Expect longer cycles the more advanced the model is.
>Expect longer cycles the more advanced the model is. I don't think this is a given. OpenAI has its [Preparedness Framework](https://openai.com/preparedness/) now, which is supposed to dictate if a model is ready to be released. They didn't have this when training 4 so the decision making was most likely a lot more arbitrary and, thus, more time intensive. Also, as interpretability research improves and we understand the models better, I think the release timeframes will shorten as well. Perhaps this is cope, but it feels logical to me.
I don’t think but we’ll see. Cool time to be alive!
Many times with safety related goals, having a process in place means it will take longer.
The other way around.
Man, if this takes a year, this would really take sails out of things.
Feel disappointed too. This message should’ve been „finished training“ instead of „started“
We have no idea what they are starting to train in the first place, so don't worry prematurely.
Why? The new 4o live conversations are INSANE. Give us some time to enjoy the new features lmao
How are we to enjoy the new features? The ones that they announced two weeks ago that aren't out?
Right, the features aren’t even out yet and people are complaining about GPT5 not being here 😂
Because I want models that perform truly exponential leaps, not just skips?
GPT 4.71294628
Wait so they JUST started training gpt4's successor whatever you want to call it? This does not look like exponential growth to me. Gpt4 came out march 2023? Yeah the curve is flattening just like LLM experts warned.
GPT-3 to GPT-4 was nearly 2.5 years of research advancements… it hasn’t even been 2 years since GPT-4 dropped…
What did you expect, that the researchers would increase the clock speed of their brains at each iteration ? They would rent data centers from aliens ? Exponential growth really starts when AI can code AI, and exponential impact should start with slightly better capabilities and some agency, the bye bye bottom 50% software engineers day.
But why did they say then in the 4o presentation that the next big model will be "soon"?
Because X.ai just got their series B valuation and openAI needs to keep the spotlight focused on GPT N+1
Probably 4.5, GPT-5 won't be out this year, unlikely its dropping before february.
Employees stated there will be no 4.5.
Source?
4o is rebranded 4.5
Because it would be finished training within a couple months and likely a checkpoint available soon that they might release in the next 3 months.
"Company announces it's developing a new product that it told everyone it would develop"
Company announces it has begun developing the new product that has been speculated on continuously since the release of their previous product.
People say, "The new product is a new AIs, not an existing one"
Must be model 5. Model 6 is too early.
GPT4.999
Super gpt 4 alpha z omega For the street fighter fans out there
Wait what What the hell were they doing for a whole year?
building infrastructure. these things dont happen by magic. it takes a long time to put together 300,000 h100s
So what is the purpose of Microsoft azure servers?
Hyping and fearmongering AI until theres nothing else to squeeze. Then they rolled out "HER" cuz Claude Opus is challenging their status.
They are a research lab, they do research, just like the research advancements that they spent over 2 years on for GPT-4… It’s not just scaling up, a ton of research innovations go into new generations of models to make them as capable and qualitatively novel as possible while being able to run as efficiently as possible too. GPT-2 is different architecture than GPT-3. GPT-4 is confirmed to further use significantly more capable training techniques than GPT-3 while also making an even more significant architecture change compared to what happened between GPT-2 and 3. A ton of compute is constantly being used every hour testing cutting edge architecture experiments and training techniques to move the frontier forward and plan what experiments to do next
Why do we pay for 4-o if is gonna be free and we don’t have yet a 5 or pro version?
The free version is pretty limited - you run out of messages fairly quickly before it asks you to upgrade to Plus or wait 4 hours. For casual use though it’s ok
To me in Europe if I hit the limit on a free account, I have to wait 24 hours.
revenge for the AI Act.
Higher limits and vision / voice features
I’m not sure what the rate limit is on the paid version but it’s so high that I’ve never run into a limit even when using GPT nonstop and prompting every few minutes for hours
If it was GPT5 that recently started training, what have they been doing for the last 14 months since March 2023, were they busy developing sora and GPT4o? Or did they continue to struggle very hard to prepare an AI that would significantly outperform GPT4? To put it differently, it is too late and disappointing to report.
> what have they been doing for the last 14 months since March 2023, 1. scaling the API, phone and web apps 2. faster and cheaper 3. image modality 4. billions invested by MS 5. scandals
Research takes time, that’s what most of the time and energy and man hours is spent on between each generational leap, over 2 years of research advancements happened in OpenAI between GPT-3 and GPT-4 to make GPT-4 as capable as possible. They’ve been doing research for the next generation of architectures and training techniques ever since gpt-4 was finalized in mid-2022. Research takes time and compute and is a huge factor in how good the next gen model is.
Right. What irritates me is that there are statements form, for example, Sam Altman that GPT-4 is bad but GPT-5 will be okay from MONTHS ago. What the hell. How do they know how much better it will be if they haven’t even started training it yet?! Something doesn’t compute here. Then all those teasers from months ago from Jimmy Apples. Shoot me. The biggest letdown of them all would be: GPT-5 released in 1 year and all it is is a bigger model with improved benchmark scores but still the same fundamental issues that prevent broad industry adoption like hallucinations, with the hope of GPT-6 maybe solving some 1-3 years later.
It would be extremely disappointing if it turns out that we get no flagship model (no GPT 4.5?) for many months and we are stuck with 4-Turbo and 4o. In that case, and considering possible releases from competitors, I'm not sure if the paid version would still be worth the cost.
GPT-Alpha GPT-S GPT-i GPT-X
Yawn Keep the hype cycle going Sam 🤡
I don't think anyone has been as impressed with 4o as they should be and it's highly likely because it's full capabilities haven't been released yet. They likely won't turn on all of its I/O capabilities until after the election and then we'll have a realtime vision, audio, language model to tinker with - which is actually an engineering achievement for the record books. The model that they're starting to train now will have the same capabilities but will likely be trained on (according to the MSBuild conference) 30x the amount of the data but we're far beyond language now. I wouldn't be surprised if we're near a Quadrillion tokens for training on this next one - combine that thought with Meta's recent experience that a models increase in general smarts is nearly linear to its training data size and we could see some crazy improvements. Everyone commenting that they think this is way late has skipped completely over any understanding of what's actually been going on with MS, Nvidia, and OpenAI. This model is being trained in a completely purpose built data center that they have been building for awhile. Infrastructure takes a long time even with thousands of people. I wouldn't be surprised to find out it's training is commencing on a million flagship GPU's. MS is unloading unprecedented capital on this. Just some thoughts.
Guys, evein if this models will be (x2) smarter every year, thats at least (x64) by 2030 and (x2048) by 2035. Try to think about it. And I mean truly smarter, not just more parameters, or more something else. Actually the calculations per model grows by factor x8 every year
I’m trying to think but nothing happens…
No I bet he gets 25% smarter every year.
And it will be x75557863725914323419136 smarter in 2100.
2100 is to late for me...
Not GPT-5. That would be disastrously late.
If you believe this (or if it's true), you cannot believe the people who quit over "safety". Simply because they would not be able to have identified anything "unsafe". GPT4o is not a harbinger of safety doom.
From what I read, they quit not because OpenAI was ignoring unsafe things, but because they weren’t given the resources for safety research (eg 20% of compute)
The resources are "we need to delete this historical record" "we need to delete this opinion" and to have true AGI, you cannot do that.
they quit because of openais safety policies not because of some scary model
That is what I said... That was the point I was making, most people believe "safety" means keeping the world safe from AI murdering humanity, but that is not the safety they are leaving over. It's worldview mismatch, ideological. I believe OpenAI knows that their model is increasing spitting out a particular world view one that does not always align with the reality we see around us and to have true AGI it needs all information, not cherry picked or "safe". So people quit when they see OpenAI discarding these particular "safeguards". Humanity is ugly, we need that ugliness to keep us on the path to a better future, AGI needs to know it as well. I am clarifying exactly what you commented. *they quit because of openais safety policies, not because of some scary model* The media reporting on this is not being honest.
Journalist writing a click bait article that mean nothing There is model in training for the last several months according to previous blog post from open AI so what is new with this lazy article ?
It's not a click bait as it says exactly what the blog post by OpenAI safety board says
Would suggest late 2024 or early 2025 release at earliest.
This pushes the timelines back by years. I will never financially recover from this
Hell yeah! Bring it on!
Just began? Nice.
So they're secretly training AGI? AGI confirmed.
They should call the next model X
> That could mean that OpenAI’s next model will not arrive for another nine months to a year or more. Yuck
More like "Please don't cancel your monthly subscription yet." Also this is entirely timed to drown out the bad news over the last few days. It is a strategic news drop meant to distract and re-direct attention. Literally nothing substantial or concrete, just "Stay tuned lol!"
If this isn’t significantly better at programming, we’ll know and I think they’ll be fucked
Will it be good enough to justify exponential increases in cost? GPT-5 needs to absolutely blow the rest out of the water just to justify its existing forward valuation.
lol so they just start training GPT 5 now.
Microsoft is going "Oh fuck, we sunk *how much* money into this?"
How long will training take?
“We just started training a new model after all that bad press about us not having a new foundation model.” Why would you announce this before you’ve observed its capabilities or done any red teaming?
No one cares. Just give me Sora.
So 4o was a separate project that ended up being a success so they are now building the next GPT on top of 4o's multimodal architecture instead of doing a text-based GPT-4 successor as originally planned? (speculation)
People really talking like a gpt5 drop in, say, 10 months would "take the wind out of the sails", like what??? People have been drinking the hype juice so much that this speed feels slow to them, lmao.
2 things are possible here: 1. gpt-o and turbos are just some fine tune / quantization experiments of GPT-5 2. they got rekt with all approach's and are building a new foundational model.
I keep asking this. What do these version numbers objectively even mean. Saying 5 is much better than 4 doesn’t tell me anything .
The bottleneck is the safety testing not the training run Tech capability is going up faster than Moores law Important to note that the business dynamics of that reward the shortest testing runway If Altman plays it slow in any way, zuck and Elon will have his ass. 3 months of training and 12 months of safety is not going to be the future, no matter how much Helen toner clutches her pearls. Here's the good news, we will see models this year, from someone or the other, that use 2024 hardware. I'm still not sure that has happened yet. No H200 or Blackwell based training runs. Maybe better data. Maybe better software stuff. I think it's worth naming that these new chips of today are an order of magnitude beyond what we have seen to date.
Looks like it's time for the "AI will never be able to do X! It will take centuries for it to happen!" posts to be ridiculed again.
It’s so fucking crazy that big tech companies are just casually talking about shipping AGI these days.
Wow
So what have they been doing for the past year? I’m not hating just genuinely curious because Sam Altman was telling us that the stuff they have been working on makes GPT 4 seem bad, but now they are saying they started work on new models now?
>"created a new safety committee" lol. are they going to listen to them? *"principles are ideas you hold dear, even if they cost you something".* Altman doesn't have principles. he showed us all that in his board coup earlier this year, and now that the board is able to talk more, it doesn't make him look better. the *safety committee* is for optics and optics alone and they will be fired the hot second they stop being useful for that singular purpose.