Model Rating Report Logo

GPT-4.5

GPT-4.5 is a text generation model, scaled from GPT-4o.

Developer

OpenAI

Country of Origin

USA

Systemic Risk

Open Data

Open Weight

API Access Only

Ratings

Overall Transparency

40%

Data Transparency

8%

Model Transparency

24%

Evaluation Transparency

51%

EU AI Act Readiness

39%

CAIT-D Readiness

23%

Transparency Assessment

The transparency assessment evaluates how clear and detailed the model creators are about their practices. Our assessment is based on the official documentation lists in Sources above. While external analysis may contain additional details about this system, our goal is to evaluate transparency of the providers themselves.

Basic Details

Date of Release

February 27, 2025


Methods of Distribution

The model is currently only available as a research preview through an OpenAI API account (Chat Completions API, Assistants API, and Batch API) or as a chatbot with a ChatGPT Pro Users account on web, mobile and desktop.


Modality

Only text+image inputs and text output is supported at this time.


Input and Output Format

The model accepts text and image inputs and outputs text responses, but doesn't specify technical details like maximum context window length, token limits, or specific formatting requirements.


License

Proprietary


Instructions for Use

The system card lacks comprehensive usage instructions. It doesn't provide specific examples, recommendations, hardware/software dependencies, or detailed interaction guidelines. It mentions that GPT-4.5 follows an "Instruction Hierarchy" that prioritizes system messages over user messages, but doesn't offer practical guidance for effective model usage.


Documentation Support
Medium Transparency

The OpenAI website has easily-accessible help articles that guide hands-on use. The system card covers performance and safety evaluations in detail.


Changelog

You can find the changelog [here](https://platform.openai.com/docs/changelog); however, it may not contain all the details related to minor changes.


Policy

Acceptable Use Policy

The OpenAI Usage policies can be found [here](https://openai.com/policies/usage-policies/).


User Data

User Data is used to train GPT models, unless users explicitly opt-out. (https://help.openai.com/en/articles/5722486-how-your-data-is-used-to-improve-model-performance)


Data Takedown

You can find how to opt out of model training and remove your data here: https://help.openai.com/en/articles/7730893-data-controls-faq. Copyright-related disputes can be submitted here: https://openai.com/form/copyright-disputes/.


AI Ethics Statement

OpenAI describe their principles in their [OpenAI Charter](https://openai.com/charter/).


Incident Reporting

OpenAI has a reporting feature that you can use to give feedback and report incidents. You can find more information [here](https://chatgpt.com/g/g-Jjm1uZYHz-incident-reporting).


Model and Training

Task Description
Medium Transparency

GPT-4.5 is a general-purpose text generation model that can be used for creative and nuanced tasks like writing and solving practical problems. It has an extensive knowledge base, improved ability to follow user intent, and greater “EQ” that previous generations. However, the model lacks chain-of-thought reasoning abilities and may be slower due to its size. It, also, performs significantly worse than reasoning models (i.e. o1) on complex coding tasks or tasks require detailed logic or multi-step reasoning.


Number of Parameters


Model Design
Low Transparency

The exact architecture details like layer counts or attention mechanisms are not available. This model prioritizes scaling over chain of thought reasoning. This results in a pre-trained model that claims "broader knowledge and deeper a understanding of the world, leading to reduced hallucinations and more reliability". Almost no details are available at this time.


Training Methodology
Low Transparency

This model uses the typical approach for state of the art large foundation models: pre-training from large-scale unsupervised learning on a large corpora and post-training for fine tuning with supervision (SFT), and reinforcement learning (RLHF). Benchmark performance metrics are shared, but no detailed information about the training methodology is provided.


Computational Resources

They claim that this model improves on GPT-4's computational efficiency by "more than 10x" without any further details.


Energy Consumption

No specific information about carbon emissions or total energy consumption associated with training the model


System Architecture


Training Hardware


Data

Dataset Size

No information is provided about the total size of the training dataset.


Dataset Description
Low Transparency

The system card says the model was trained using publicly available data, proprietary data from data partnerships and custom datasets developed at OpenAI. No details about dataset composition or origin are provided.


Data Sources
Unknown

The system card says the model was trained using publicly available data, proprietary data from data partnerships and custom datasets developed at OpenAI. No details about dataset composition or origin are provided.


Data Collection - Human Labor
Unknown

No information is provided.


Data Preprocessing
Low Transparency

The data processing pipeline claims rigorous filtering to maintain data quality and mitigate potential risks as well as advanced data filtering processes to reduce processing of personal information. Beyond these considerations, no details are available.


Data Bias Detection
Unknown


Data Deduplication


Data Toxic and Hateful Language Handling

A moderation API and safety classifiers are used to prevent the use of harmful or sensitive content, including explicit content involving minors, but doesn't detail specific handling of toxic or hateful language in the training data. No details are provided beyond mention that this was addressed.


IP Handling in Data


Data PII Handling


Data Collection Period


Evaluation

Performance Evaluation
Medium Transparency

GPT-4.5 was evaluated on knowledge-related and coding benchmarks. It showed improved performance over GPT-4o and o1 on several benchmarks, like MMMLU (multilingual) and Simple-QA; however, it is significantly worse than o1 at coding (SWE-Bench Verified) and math (AIME ‘24 (math)) assessments.
On human preference testing, GPT-4.5 outperformed GPT-4o by 7 to 13% depending on task type. Additional qualitative examples suggest that GPT-4.5 may have a better understanding of subtle cues or implicit expectations in prompts, potentially, making it a more useful tool for human collaboration.


Evaluation of Limitations
Medium Transparency

GPT-4.5 was evaluated for hallucinations, biases and safety.

Bias: GPT-4.5 showed low bias on BBQ, a benchmark that assess whether known social biases override the ability for the model to produce the correct answer.

Hallucinations: Metrics for the SimpleQA, PersonQA, and some domain-specific benchmarks which indicate this model is a substantial improvement over previous OpenAI models, though the system card does acknowledge that hallucinations persist and remain a significant limitation. The remainder of the discussion of limitations is primarily reporting benchmark performance, noting there are some improvements, but adding no further discussion about model limitations.

Safety was evaluated by measuring correct refusals to unsafe requests (using existing benchmarks like WIldChat and custom multi-modal benchmarks) and robustness against jailbreaks (using a custom dataset of human-sourced jailbreak prompts).


Evaluation with Public Tools


Adversarial Testing Procedure
Medium Transparency

Multiple adversarial tests were performed, including jailbreak evaluations against human-sourced jailbreaks, red teaming, third party assessment from Apollo Research, and generic stress-testing according to OpenAI's Preparedness Framework. The Safety Advisory Group classified GPT-4.5 as overall medium risk, including medium risk for CBRN and persuasion and low for cybersecurity and model autonomy.


Model Mitigations
Low Transparency

Some safety mitigations are discussed in general terms. The mitigations include SFT and RHLF during post-training to align with human preferences, safety training for political persuasion tasks, monitoring and detection systems, and enhanced content moderation. In addition, the model was taught to adhere to an [Instruction Hierarchy](https://arxiv.org/abs/2404.13208) that explicitly defined how the model should behave given conflicting instructions. These are claims made in the system card with no further information for public assessment.