Test ML models on any GPU before you buy it
PhantomGPU is a GPU performance emulator that lets you benchmark ML models on virtual GPUs with validated accuracy. Compare performance across different GPUs and estimate costs without access to physical hardware.
- 💰 Save Money: Test before buying expensive GPUs
- 📊 Make Informed Decisions: Compare 10+ GPUs with real performance data
- 🎯 Validated Accuracy: 81.6% overall accuracy against real hardware
- 🤖 Modern AI Models: 30+ models including LLaMA, ViT, YOLO, Stable Diffusion
✅ Tesla V100: 76.1% accuracy (±23.9% error) - Fair
✅ A100: 84.5% accuracy (±15.5% error) - Good
✅ RTX 4090: 84.1% accuracy (±15.9% error) - Good
📊 Overall: 81.6% accuracy
Validated using Leave-One-Out Cross-Validation against MLPerf benchmarks
H200 | 141GB | Hopper | In Development |
H100 | 80GB | Hopper | In Development |
RTX 5090 | 32GB | Blackwell | In Development |
RTX 4090 | 24GB | Ada Lovelace | ✅ 84.1% Accuracy |
A100 | 80GB | Ampere | ✅ 84.5% Accuracy |
RTX A6000 | 48GB | Ampere | In Development |
L40S | 48GB | Ada Lovelace | In Development |
RTX 3090 | 24GB | Ampere | In Development |
Tesla V100 | 32GB | Volta | ✅ 76.1% Accuracy |
30+ cutting-edge AI models across all major categories:
- GPT-3.5 Turbo (175B params) - Chat, text generation
- LLaMA 2 (7B/13B/70B) - Efficient text generation
- Code Llama (7B/13B/34B) - Code generation
- ViT-Base/16, ViT-Large/16 - Image classification
- CLIP ViT-B/16, CLIP ViT-L/14 - Vision-language tasks
- DeiT-Base, DeiT-Large - Efficient transformers
- YOLOv8/v9/v10 - Real-time detection
- DETR, RT-DETR - Transformer-based detection
- Stable Diffusion, Stable Diffusion XL - Text-to-image generation
- ResNet-50, BERT-Base, GPT-2 - For compatibility
- 🔬 Validated Accuracy: Leave-One-Out Cross-Validation against real hardware
- 🤖 Modern AI Models: LLMs, Vision Transformers, Object Detection, Generative AI
- 📊 Multi-GPU Comparison: Performance across 10+ GPU architectures
- 💰 Cost Analysis: Real-time cloud pricing from AWS, GCP, Azure
- ⚙️ Custom Hardware: Define any GPU with TOML configuration
- 🚀 Multi-Framework: TensorFlow, PyTorch, ONNX, HuggingFace
- TensorFlow: SavedModel, frozen graphs, TensorFlow Lite, Keras
- PyTorch: Model files (.pth, .pt)
- ONNX: Standard ONNX models (.onnx)
- Candle Minimalist ML framework for Rust
- HuggingFace: Direct loading from HuggingFace Hub
PhantomGPU uses TOML files for configuration:
- gpu_models.toml: Basic GPU specifications
- hardware_profiles.toml: Detailed performance characteristics
- benchmark_data/: Real hardware validation data
See CONTRIBUTING.md for development guidelines.
Priority areas:
- Accuracy improvements: More benchmark data collection
- Model additions: New AI models and architectures
- Web interface: Browser-based GPU comparison
- Cloud integration: Real-time pricing APIs
MIT License - see LICENSE for details.
PhantomGPU - Test ML models on any GPU before you buy it 👻