- Shanghai, China
Lists (11)
Sort Name ascending (A-Z)
Stars
- All languages
- ApacheConf
- Arduino
- Assembly
- Astro
- Awk
- C
- C#
- C++
- CSS
- Cirru
- Clojure
- CoffeeScript
- Common Lisp
- Crystal
- Cuda
- D
- Dart
- Dockerfile
- Eagle
- Elixir
- Elm
- Emacs Lisp
- Frege
- GCC Machine Description
- GLSL
- Go
- Groovy
- HTML
- Haskell
- IDL
- Java
- JavaScript
- Julia
- Jupyter Notebook
- Kotlin
- LLVM
- Lua
- MATLAB
- MDX
- Makefile
- Markdown
- Nix
- Nunjucks
- Objective-C
- Objective-C++
- OpenEdge ABL
- PHP
- PowerShell
- Processing
- Python
- R
- Rich Text Format
- Roff
- Ruby
- Rust
- SCSS
- Scala
- Shell
- SourcePawn
- Swift
- TeX
- TypeScript
- V
- Vala
- Vim Script
- Vue
- XML
Implementation of papers in 100 lines of code.
Detail code implementation and experimental setting for our paper: Federated Learning on Multilabel Evolving Data Streams
What are the principles we can use to build LLM-powered software that is actually good enough to put in the hands of production customers?
PyTorch Wildlife: a Collaborative Deep Learning Framework for Conservation.
The Startup CTO's Handbook, a book covering leadership, management and technical topics for leaders of software engineering teams
An open-source framework for machine learning and other computations on decentralized data.
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
DeepSeek LLM: Let there be answers
FlashMLA: Efficient Multi-head Latent Attention Kernels
Analyze computation-communication overlap in V3/R1.
A lightweight data processing framework built on DuckDB and 3FS.
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
A bidirectional pipeline parallelism algorithm for computation-communication overlap in DeepSeek V3/R1 training.
An elegant PyTorch deep reinforcement learning library.
A high-performance distributed file system designed to address the challenges of AI training and inference workloads.
DeepSeek-VL2: Mixture-of-Experts Vision-Language Models for Advanced Multimodal Understanding
Integrate the DeepSeek API into popular software
This repository contains the source code for the Saving 77% of the Parameters in Large Language Models Technical Report
MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.
Keep searching, reading webpages, reasoning until it finds the answer (or exceeding the token budget)
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. 🔥 [Paper + Code + Demo]
Machine Learning Engineering Open Book


