Skip to content
View bzantium's full-sized avatar

Block or report bzantium

Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
bzantium/README.md

Minho Ryu (bzantium)

Senior AI Research Engineer | Foundation Models & Scalable Systems | Google Developer Expert (AI)

Linkedin Badge Gmail Badge Google Scholar Badge

πŸ‘‹ About Me

I am a Senior AI Research Engineer and Google Developer Expert (AI) with a mission to architect and scale the powerful, efficient, and accessible large language models that will define the future.

My expertise covers the full lifecycle of foundation models: from curating massive datasets and architecting cutting-edge training infrastructure to developing production-grade models that set new performance benchmarks. I thrive on solving complex, large-scale challenges and am deeply invested in strengthening the open-source ecosystem that fuels global AI innovation.

πŸš€ Key Professional Highlights

  • Foundation Model Development: Co-led the end-to-end pre-training of Kakao's Kanana V1 foundation model from a 3T token dataset and implemented compute-efficient scaling techniques like Pruning & Distillation. I also spearheaded key enhancements for Kanana-1.5 (including its 128K long-context extension) and owned the full development of a production embedding model that surpassed larger competitors.
  • Scalable AI Infrastructure: Architected and optimized a cutting-edge, scalable LLM training pipeline from the ground up using JAX, MaxText, and TPUs. This work was featured in an official Google Cloud Blog Post and my expertise was recognized with a presentation at Google Cloud Next 2025 (YouTube).
  • Open Source Leadership: As a Research Lead at EleutherAI, I co-led the development of Polyglot-Ko, the first open-source Korean large language model, successfully training and releasing models up to 12.8B parameters.

πŸ™ Open Source Contributions

My GitHub activity reflects a consistent track record of contributing high-impact code to the core of the modern AI ecosystem. I focus on strengthening foundational libraries, building scalable systems, and advancing rigorous evaluation. Below are some of my key contributions:

✍️ Publications

Pinned Loading

  1. EleutherAI/polyglot EleutherAI/polyglot Public

    Polyglot: Large Language Models of Well-balanced Competence in Multi-languages

    484 42

  2. lassl/lassl lassl/lassl Public

    Easy Language Model Pretraining leveraging Huggingface's Transformers and Datasets

    Python 130 15

  3. EleutherAI/oslo EleutherAI/oslo Public

    OSLO: Open Source for Large-scale Optimization

    Python 175 26

  4. AI-Hypercomputer/maxtext AI-Hypercomputer/maxtext Public

    A simple, performant and scalable Jax LLM!

    Python 2.2k 493

  5. kakao/kanana kakao/kanana Public

    Kanana: Compute-efficient Bilingual Language Models

    279 15

  6. kakao/kanana-2 kakao/kanana-2 Public

    22