AI Model Benchmarks
Compare performance metrics across different AI models and tools
Language Models
Performance metrics for large language models
Model | Generation Speed(tokens/s) | Context Length(tokens) | Cost($/1K tokens) | Memory Usage(GB) | Last Updated |
---|---|---|---|---|---|
![]() GPT-4 Turbo OpenAI | 120 | 128000 | 0.01 | 16 | 2024-03-15 |
![]() Claude 3 Opus Anthropic | 145 | 200000 | 0.015 | 24 | 2024-03-10 |
Image Generation
Benchmarks for image generation models
Model | Generation Time(seconds) | Max Resolution(pixels) | Cost($/image) | CLIP Score(score) | Last Updated |
---|---|---|---|---|---|
![]() DALL-E 3 OpenAI | 3.5 | 1024x1024 | 0.04 | 0.92 | 2024-03-15 |
![]() Midjourney V6 Midjourney | 4.2 | 1024x1024 | 0.05 | 0.89 | 2024-03-12 |
Speech Models
Performance metrics for speech recognition and synthesis
Model | Word Error Rate(%) | Real-time Factor(x) | Latency(ms) | Languages(count) | Last Updated |
---|---|---|---|---|---|
![]() Whisper Large v3 OpenAI | 2.8 | 0.5 | 250 | 100 | 2024-03-14 |
![]() AssemblyAI AssemblyAI | 3.2 | 0.6 | 300 | 85 | 2024-03-10 |
Code Generation
Benchmarks for code generation and completion models
Model | Accuracy(%) | Response Time(ms) | Languages(count) | Test Pass Rate(%) | Last Updated |
---|---|---|---|---|---|
![]() GitHub Copilot GitHub | 85 | 150 | 40 | 78 | 2024-03-15 |
![]() Amazon CodeWhisperer Amazon | 82 | 180 | 35 | 75 | 2024-03-12 |
![]() Tabnine Tabnine | 80 | 120 | 30 | 72 | 2024-03-08 |
Enterprise Metrics
Performance and reliability metrics for enterprise AI services
Model | Uptime(%) | Rate Limit(req/min) | SLA(%) | Support Time(hours) | Last Updated |
---|---|---|---|---|---|
![]() Azure OpenAI Microsoft | 99.99 | 10000 | 99.9 | 1 | 2024-03-15 |
![]() AWS Bedrock Amazon | 99.95 | 8000 | 99.9 | 2 | 2024-03-14 |
![]() Vertex AI Google Cloud | 99.95 | 9000 | 99.9 | 1.5 | 2024-03-13 |