Train AI models from scratch. Join the distributed training network. Contribute GPU power to build something amazing together.
Train 1M, 2.7M, 10M, 25M, or 50M parameter models. Pick what fits your device.
Watch training in real-time. See loss curves, speed, and generated samples.
Talk to your trained model with a web UI. Includes text-to-speech.
Join the Hive network. Train across multiple machines and GPUs.
Native CUDA support for NVIDIA GPUs. Also works on Apple Silicon.
Understand transformers by building one from scratch. Full source code.
Works on GPU, CPU, low-VRAM devices, Colab, and even Android via Termux.
Each peer trains the model on their own GPU/CPU
Send learning updates to the coordinator
Coordinator combines gradients from all peers
Updated model is broadcast back to all peers
๐ More peers = faster training! Each GPU adds compute power. An RTX 4090 + Colab T4 together train ~10% faster than alone.
Contribute your GPU power to distributed AI training