Useful information

Prime News delivers timely, accurate news and insights on global events, politics, business, and technology

OpenAI confirms the new o3 and o3-mini frontier models


Join our daily and weekly newsletters to get the latest updates and exclusive content on industry-leading AI coverage. More information


OpenAI is slowly inviting select users to try out a brand new set of reasoning models called o3 and o3 mini, successors to the o1 and o1 mini models that were just fully released earlier this month.

OpenAI o3, named so to avoid copyright issues with phone company O2 and because CEO Sam Altman says the company “has a tradition of being really bad with names,” consists of two models, o3-mini and o3. Altman announced the o3 series today on the final day of OpenAI’s “12 Days of OpenAI” livestreams and said they would initially be handed over to outside researchers to security testing.

“We see this as the beginning of the next phase of AI, where you can use these models to perform increasingly complex tasks that require a lot of reasoning,” Altman said. “For the last day of this event we thought it would be fun to go from one frontier model to the next frontier model.”

Altman said during the livestream that the company plans to launch o3-mini in late January and o3 “shortly after.”

Altman also said that the o3 model was “amazing at coding” and the benchmarks shared by OpenAI back this up, outperforming even the o1 in programming tasks.

Outstanding Encoding Performance: o3 outperforms o1 by 22.8 percentage points on SWE-Bench Verified and achieves a Codeforces score of 2727, surpassing OpenAI’s chief scientist score of 2665.

Mathematics and Science Mastery: o3 scores 96.7% on the AIME 2024 exam, missing only one question, and achieves 87.7% on GPQA Diamond, far surpassing expert human performance.

Border reference points: The model sets new records on challenging tests like EpochAI’s Frontier Math, solving 25.2% of problems where no other model surpasses 2%. In the ARC-AGI test, o3 triples the score of o1 and exceeds 85% (as verified live by the ARC Prize team), representing a milestone in conceptual reasoning.

Deliberative alignment

Along with these advancements, OpenAI reinforced its commitment to security and alignment.

The company presented New research on deliberative alignmenta fundamental technique for creating your most robust and aligned model to date.

This technique embeds human-written security specifications into models, allowing them to explicitly reason about these policies before generating responses.

The strategy seeks to solve common security challenges in LLMs, such as vulnerability to jailbreak attacks and over-rejection of benign prompts, by equipping the models with chain-of-thought (CoT) reasoning. This process allows models to remember and apply security specifications dynamically during inference.

Deliberative alignment improves on previous methods such as reinforcement learning from human feedback (RLHF) and constitutional AI, which rely on security specifications only for label generation instead of incorporating policies directly into the models.

By fine-tuning LLMs based on security-related cues and their associated specifications, this approach creates models capable of policy-based reasoning without relying heavily on human-labeled data.

Results shared by OpenAI researchers in a new article, not peer reviewed indicate that this method improves performance on security benchmarks, reduces harmful outcomes, and ensures better adherence to content and style guidelines.

Key findings highlight the advancements of the o1 model over its predecessors such as GPT-4o and other next-generation models. Deliberative alignment allows the o1 series to excel at resisting jailbreaks and providing secure completions while minimizing excessive rejections on benign prompts. Furthermore, the method facilitates out-of-distribution generalization, showing robustness in encrypted and multilingual jailbreak scenarios. These improvements align with OpenAI’s goal of making AI systems more secure and interpretable as their capabilities grow.

This research will also play a key role in the alignment of o3 and o3-mini, ensuring their capabilities are powerful and responsive.

How to request access to the o3 and o3-mini test

Early access applications are now open at the AI Open Website and will close on January 10, 2025.

Applicants must fill out an online form which asks them for a variety of different pieces of information, including links to previously published articles and their code repositories on Github, and they select which of the models (o3 or o3-mini) they want to try, as well as what they plan. to use them.

Selected researchers will have access to o3 and o3-mini to explore their capabilities and contribute to security assessments, although the OpenAI form warns that o3 will not be available for several weeks.

Researchers are encouraged to develop robust evaluations, create controlled demonstrations of high-risk capabilities, and test models in scenarios not possible with widely adopted tools.

This initiative builds on the company’s established practices, including rigorous internal security testing, collaborations with organizations such as the US and UK AI Safety Institutes, and its Readiness Framework.

The application process asks for details such as research focus, past experience, and links to previous work. OpenAI will review applications on a rolling basis and selections will begin immediately.

A new leap forward?

The introduction of o3 and o3-mini signals a leap forward in AI performance, particularly in areas that require advanced reasoning and problem-solving capabilities.

With their exceptional results in coding, mathematics, and conceptual benchmarks, these models highlight the rapid progress being made in AI research.

By inviting the broader research community to collaborate on security testing, OpenAI aims to ensure that these capabilities are implemented responsibly.

Watch the broadcast below:



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *