Skip to content

open-gigaai/giga-train

GigaTrain

An Efficient and Scalable Training Framework for AI Models

| Quick Start | Contributing | License | Citation |

✨ Introduction

GigaTrain is an efficient and scalable training framework engineered to accelerate the development of large AI models. It provides optimized performance and streamlined training workflows, allowing researchers and developers to easily experiment with various models.

Major features
  • 🔍 Unified distributed training: Seamless multi-GPU/multi-node execution; supports DeepSpeed ZeRO (0/1/2/3), FSDP/FSDP2, DDP, etc.
  • 🔧 Flexible and reproducible configs: Clean PY/YAML/JSON configuration and a registry-driven, modular design with pluggable optimizers, schedulers, samplers, transforms, etc.
  • 📈 Performance and memory efficiency: Mixed precision (FP16/BF16/FP8), gradient accumulation, gradient checkpointing, EMA, etc.
  • 📊 Built-in monitoring and checkpointing: Integrated experiment logging and robust checkpointing for reliable long runs and resumability.
  • Lightweight and Easy to Use: Simple pip/source install; developers can focus solely on implementing the key algorithm, as the framework handles repetitive, tedious, and error-prone things like backprop, logging, checkpointing, resuming, EMA, and multi-node/multi-GPU execution.

⚡ Installation

GigaTrain can be installed from PyPi and has to be installed in a virtual environment (venv or conda for instance):

pip3 install giga-train

or you can install directly from source for the latest updates:

conda create -n giga_train python=3.11.10 conda activate giga_train git clone https://github.com/open-gigaai/giga-train.git cd giga-train pip3 install -e .

🚀 Getting Started

  • We provide a step-by-step example to teach you how to fine-tune a model using GigaTrain.
  • Refer to projects for more examples.

🤝 Contributing

We welcome contributions! Please see our Contributing Guide for details.

📄 License

This project is licensed under the Apache License 2.0 - see the LICENSE file for details.

📖 Citation

@misc{gigaai2025gigatrain, author = {GigaAI}, title = {GigaTrain: An Efficient and Scalable Training Framework for AI Models}, year = {2025}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\url{https://github.com/open-gigaai/giga-train}} }

About

GigaTrain: An Efficient and Scalable Training Framework for AI Models

Resources

License

Code of conduct

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages