aitoolkit.co logo
aitoolkit.co

AI Model Benchmarks

Compare performance metrics across different AI models and tools

Language Models

Performance metrics for large language models

Model
Generation Speed(tokens/s)
Context Length(tokens)
Cost($/1K tokens)
Memory Usage(GB)
Last Updated
GPT-4 Turbo
GPT-4 Turbo
OpenAI
120
128000
0.01
16
2024-03-15
Claude 3 Opus
Claude 3 Opus
Anthropic
145
200000
0.015
24
2024-03-10

Image Generation

Benchmarks for image generation models

Model
Generation Time(seconds)
Max Resolution(pixels)
Cost($/image)
CLIP Score(score)
Last Updated
DALL-E 3
DALL-E 3
OpenAI
3.5
1024x1024
0.04
0.92
2024-03-15
Midjourney V6
Midjourney V6
Midjourney
4.2
1024x1024
0.05
0.89
2024-03-12

Speech Models

Performance metrics for speech recognition and synthesis

Model
Word Error Rate(%)
Real-time Factor(x)
Latency(ms)
Languages(count)
Last Updated
Whisper Large v3
Whisper Large v3
OpenAI
2.8
0.5
250
100
2024-03-14
AssemblyAI
AssemblyAI
AssemblyAI
3.2
0.6
300
85
2024-03-10

Code Generation

Benchmarks for code generation and completion models

Model
Accuracy(%)
Response Time(ms)
Languages(count)
Test Pass Rate(%)
Last Updated
GitHub Copilot
GitHub Copilot
GitHub
85
150
40
78
2024-03-15
Amazon CodeWhisperer
Amazon CodeWhisperer
Amazon
82
180
35
75
2024-03-12
Tabnine
Tabnine
Tabnine
80
120
30
72
2024-03-08

Enterprise Metrics

Performance and reliability metrics for enterprise AI services

Model
Uptime(%)
Rate Limit(req/min)
SLA(%)
Support Time(hours)
Last Updated
Azure OpenAI
Azure OpenAI
Microsoft
99.99
10000
99.9
1
2024-03-15
AWS Bedrock
AWS Bedrock
Amazon
99.95
8000
99.9
2
2024-03-14
Vertex AI
Vertex AI
Google Cloud
99.95
9000
99.9
1.5
2024-03-13