Stars
NVMeVirt: A Versatile Software-defined Virtual NVMe Device
ByteCheckpoint: An Unified Checkpointing Library for LFMs
Persist and reuse KV Cache to speedup your LLM.
A high-performance Python-based I/O system for large (and small) deep learning problems, with strong support for PyTorch.
Kubernetes Operator, Helm Charts, Ansible Playbooks, and utility scripts for large-scale AIStore deployments on Kubernetes.
Large language model fine-tuning capabilities based on cloud native and distributed computing.
Large World Model -- Modeling Text and Video with Millions Context
PyTorch code and models for V-JEPA self-supervised learning from video.
Official implementation for the paper: "Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering""
Lustre Monitoring System based on Collectd, Grafana and Influxdb
This repository is established to store personal notes and annotated papers during daily research.
Delivers efficient, stable, and secure data distribution and acceleration powered by P2P technology, with an optional content‑addressable filesystem that accelerates OCI container launch.
Pretrain, finetune and serve LLMs on Intel platforms with Ray
🏭
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
The Fastest Distributed Database for Transactional, Analytical, and AI Workloads.
Training and serving large-scale neural networks with auto parallelization.
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
CodeGeeX2: A More Powerful Multilingual Code Generation Model
Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and…
A fast GPU memory copy library based on NVIDIA GPUDirect RDMA technology
