Skip to content
View bzantium's full-sized avatar

Block or report bzantium

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
bzantium/README.md

Minho Ryu (bzantium)

Senior AI Research Engineer | Foundation Models & Scalable Systems | Google Developer Expert (AI)

Linkedin Badge Gmail Badge Google Scholar Badge

πŸ‘‹ About Me

I am a Senior AI Research Engineer and Google Developer Expert (AI) with a mission to architect and scale the powerful, efficient, and accessible large language models that will define the future.

My expertise covers the full lifecycle of foundation models: from curating massive datasets and architecting cutting-edge training infrastructure to developing production-grade models that set new performance benchmarks. I thrive on solving complex, large-scale challenges and am deeply invested in strengthening the open-source ecosystem that fuels global AI innovation.

πŸš€ Key Professional Highlights

  • Foundation Model Development: Co-led the end-to-end pre-training of Kakao's Kanana V1 foundation model from a 3T token dataset and implemented compute-efficient scaling techniques like Pruning & Distillation. I also spearheaded key enhancements for Kanana-1.5 (including its 128K long-context extension) and owned the full development of a production embedding model that surpassed larger competitors.
  • Scalable AI Infrastructure: Architected and optimized a cutting-edge, scalable LLM training pipeline from the ground up using JAX, MaxText, and TPUs. This work was featured in an official Google Cloud Blog Post and my expertise was recognized with a presentation at Google Cloud Next 2025 (YouTube).
  • Open Source Leadership: As a Research Lead at EleutherAI, I co-led the development of Polyglot-Ko, the first open-source Korean large language model, successfully training and releasing models up to 12.8B parameters.

πŸ™ Open Source Contributions

My GitHub activity reflects a consistent track record of contributing high-impact code to the core of the modern AI ecosystem. I focus on strengthening foundational libraries, building scalable systems, and advancing rigorous evaluation. Below are some of my key contributions:

✍️ Publications

Pinned Loading

  1. EleutherAI/polyglot EleutherAI/polyglot Public

    Polyglot: Large Language Models of Well-balanced Competence in Multi-languages

    482 42

  2. lassl/lassl lassl/lassl Public

    Easy Language Model Pretraining leveraging Huggingface's Transformers and Datasets

    Python 131 15

  3. pytorch-admm-pruning pytorch-admm-pruning Public

    Prune DNN using Alternating Direction Method of Multipliers (ADMM)

    Python 99 18

  4. EleutherAI/oslo EleutherAI/oslo Public

    OSLO: Open Source for Large-scale Optimization

    Python 174 26

  5. kakao/kanana kakao/kanana Public

    Kanana: Compute-efficient Bilingual Language Models

    267 15