Command Palette

Search for a command to run...

GPT-4

OpenAI

Proprietary

Context

Release Date
Mar 14, 2023
Knowledge Cutoff
Sep 01, 2021
Window
8,192

PricingPer 1M tokens

Input
$30
Output
$60
Blended 3:1
$37.5

Capabilities

Speed
23 t/s
Input
Output
Reasoning tokens

Latency

TTFT
0.79 ms
500 token response
22.35 s

Benchmarks

Intelligence
●●○○○

GPT-4

GPT-4 Model Card

  Model Summary

GPT-4, developed by OpenAI, is a multimodal large language model succeeding GPT-3.5, capable of processing text and image inputs to generate text outputs. Released on March 14, 2023, it excels in reasoning, professional benchmarks, and nuanced instruction-following. It supports a context window of 8,192 or 32,768 tokens and is accessible via ChatGPT Plus, OpenAI’s API, and Microsoft Copilot.

  Performance Metrics

GPT-4 achieves human-level performance across diverse tasks, outperforming predecessors in reliability and creativity. Its multimodal capabilities enable image interpretation and complex task handling.

BenchmarkGPT-4 ScoreGPT-3.5 Score
Simulated Bar ExamTop 10%Bottom 10%
MMLU (Multiple-Choice)86.4%70.1%
HellaSwag95.3%85.5%
VQAv2 (Visual QA)77.2%Not applicable

  Architecture Details

Built on a transformer architecture, GPT-4 is pre-trained on a vast corpus of public and licensed text data, including web pages and code. Fine-tuning employs reinforcement learning with human and AI feedback (RLHF) for alignment. While exact parameter counts are undisclosed (estimated at 1.76 trillion), it achieves 10x computational efficiency over prior models, trained on an Azure-co-designed supercomputer.

  Training Data

The model leverages diverse sources, including public internet data, licensed datasets, and code repositories. Pre-training data is filtered to reduce inappropriate content, though some biases persist. Personal information is minimized where feasible to address privacy concerns.

  Safety Measures

OpenAI implemented robust safety protocols, informed by over 50 external experts across 45 languages and 29 countries. Key risks and mitigations include:

RiskMitigationResidual Risk
HallucinationsRLHF, data filteringLow, 19% better than GPT-3.5 (open-domain)
Harmful ContentRefusal training, content classifiersMedium, reduced from GPT-4-early
BiasDataset curation, refusal tuningMedium, ongoing evaluation
CybersecurityMalicious request refusalLow, 5% vulnerability rate

Red-teaming revealed reduced risks in autonomous replication (0% success) and illicit advice, though persuasion risks remain medium.

  Limitations Identified

GPT-4 may produce biased or factually incorrect outputs, particularly in specialized domains. Its image processing is less robust than text, and it lacks native audio capabilities. High inference costs and lack of real-time data access limit certain applications. Refusal behaviors may introduce quality-of-service disparities across demographics.

  Intended Applications

The model supports education, coding, content creation, and accessibility tasks, such as generating learning materials, assisting developers, or summarizing visual data. Its system message feature enables customization for specific tones or formats, like JSON outputs.

  Ethical Concerns

Potential misuse includes generating misinformation or biased content. Training data may reflect societal biases, requiring continuous mitigation. OpenAI emphasizes alignment with human values, but equitable access and preventing harmful applications remain challenges.

  Evaluation Process

Over 100 red-teaming evaluations assessed risks in cybersecurity, biorisk, and persuasion. Quantitative evaluations measured harmful content generation, with GPT-4-launch scoring 19–29% better than GPT-3.5 on hallucination avoidance. The Alignment Research Center (ARC) confirmed low risks of autonomous replication.

  Access Methods

Available via ChatGPT Plus, OpenAI’s API (image capabilities waitlisted), and Microsoft Copilot. SuperGrok subscriptions offer higher quotas; see https://x.ai/grok for details. API access is detailed at https://x.ai/api.

  Future Enhancements

OpenAI aims to improve multimodal processing (audio, video), reduce inference costs, and enhance real-time data integration. Research continues to minimize biases and strengthen reasoning for niche domains.