OpenAI o1 preview

Posted on Wednesday, September 25, 2024 by AUSTIN HARRIS, Global Sales

OpenAI recently announced the development of a new series of AI models designed to spend more time thinking before responding. These models can reason through complex tasks and solve harder problems than previous iterations in science, coding, and math.

OpenAI o1 preview, has been released in ChatGPT and the API

The first model in this series, named OpenAI o1-preview, has been released in ChatGPT and the API. This is a preview version, and OpenAI expects regular updates and improvements. Alongside this release, evaluations for the next update, which is currently in development, have also been included.

How it works

These models have been trained to take time to think through problems before responding, similar to human thought processes. Through this training, they learn to refine their thinking, explore different strategies, and recognize their mistakes.

In tests, the upcoming model update has demonstrated performance comparable to PhD students on challenging benchmark tasks in physics, chemistry, and biology. It has also shown exceptional capabilities in math and coding. For instance, in a qualifying exam for the International Mathematics Olympiad (IMO), the GPT-4o model correctly solved only 13% of problems, whereas the reasoning model scored 83%. Its coding abilities were assessed in contests, achieving the 89th percentile in Codeforces competitions. More details can be found in OpenAI's technical research post.

As an early model, OpenAI o1-preview does not yet include many features that enhance the ChatGPT experience, such as web browsing for information and the ability to upload files and images. However, for many common tasks, GPT-4o will be more capable in the near term.

This significant advancement in complex reasoning tasks represents a new level of AI capability, prompting OpenAI to reset the counter and name this series OpenAI o1.


OpenAI has introduced a new safety training approach

As part of the development of these new models, OpenAI has introduced a new safety training approach that leverages their reasoning capabilities to ensure adherence to safety and alignment guidelines. By reasoning about safety rules in context, these models can apply them more effectively.

One method of measuring safety involves testing how well the model adheres to its safety rules when users attempt to bypass them, a process known as "jailbreaking." In one of the most challenging jailbreaking tests, the GPT-4o model scored 22 on a scale of 0-100, while the o1-preview model achieved a score of 84. Further details are available in the system card and accompanying research post.

To align with the enhanced capabilities of these models, OpenAI has strengthened its safety initiatives, internal governance, and collaboration with federal governments. This includes rigorous testing and evaluations utilizing their Preparedness Framework, best-in-class red teaming, and board-level review processes, including oversight from the Safety & Security Committee.

To further their commitment to AI safety, OpenAI has recently formalized agreements with the U.S. and U.K. AI Safety Institutes. The organization has begun operationalizing these agreements, which include granting the institutes early access to a research version of the model. This initiative marks an important first step in establishing a framework for the research, evaluation, and testing of future models both prior to and following their public release.

These enhanced reasoning capabilities may be particularly beneficial for addressing complex problems in fields such as science, coding, and mathematics. For instance, the o1 model can assist healthcare researchers in annotating cell sequencing data, enable physicists to generate complex mathematical formulas required for quantum optics, and support developers across various disciplines in building and executing multi-step workflows.

More App Developer News

Tether QVAC SDK Powers AI Across Devices and Platforms



APAC 5G expansion to fuel 347B mobile market by 2030



How AI is causing app litter everywhere



The App Economy Is Thriving



NIKKE 3.5 anniversary update livestream coming soon



New AI tool targets early dementia detection



Jentic launch gives AI agents api access



Experts warn ai-generated health content risks misinterpretation without human oversight



Ludo.ai Unveils API and MCP Beta to Power AI Game Asset Pipelines



AccuWeather Launches ChatGPT Integration for Live Weather Updates



Stop Using Business Jargon: 5 Ways Buzzwords Damage Job Performance



IT spending rises as banks balance legacy and innovation



Tech hiring slumps as Software Developer job postings fall



AI is becoming more widespread in collaboration tools



FCC prohibits new foreign router models citing critical infrastructure risks



ChatGPT Carbon Footprint Matches 1.3 Million Cars Report Finds



Lens Launches MCP Server to Connect AI Coding Assistants with Kubernetes



Accelerating corporate ai investment returns



Enviromates tech startup launches global participation platform



Private Repository Secures the AI-driven Development Boom



UK Fintech Platform Enviromates Connects Projects Brands and Consumers



Env Zero and CloudQuery Announce Merger



How Industrial AI Is Transforming Operations in 2026



AI generated work from managers is damaging trust among employees



Foresight Secures $25M to Bridge Infrastructure Execution Gap



Copyright © 2026 by Moonbeam

Address:
1855 S Ingram Mill Rd
STE# 201
Springfield, Mo 65804

Phone: 1-844-277-3386

Fax:417-429-2935

E-Mail: contact@appdevelopermagazine.com