Skip to content
View shizhediao's full-sized avatar
🎯
Focusing
🎯
Focusing

Block or report shizhediao

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
shizhediao/README.md

Aloha. I'm Shizhe Diao 👋

I am a research scientist at NVIDIA Research. I was a visiting scholar at University of Illinois at Urbana-Champaign, working with Prof. Heng Ji. I received my Ph.D. in 2023 advised by Prof. Tong Zhang. My research focuses on developing methods to scale up post-training and reinforcement learning for large language models, pushing the frontier of more capable, general-purpose AI agents through data-centric design.

Besides programming, I have an immense interest in swimming 🏊, kayaking 🚣, windsurfing 🏄, dinghy sailing ⛵, and stand up paddling!

C3n7ral051nt4g3ncy

Pinned Loading

  1. OptimalScale/LMFlow OptimalScale/LMFlow Public

    An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

    Python 8.5k 831

  2. active-prompt active-prompt Public

    Source code for the paper "Active Prompting with Chain-of-Thought for Large Language Models"

    Python 250 30

  3. R-Tuning R-Tuning Public

    [NAACL 2024 Outstanding Paper] Source code for the NAACL 2024 paper entitled "R-Tuning: Instructing Large Language Models to Say 'I Don't Know'"

    Python 131 13

  4. automate-cot automate-cot Public

    Source code for the paper "Automatic Prompt Augmentation and Selection with Chain-of-Thought from Labeled Data"

    20 1

  5. T-DNA T-DNA Public

    Source code for the ACL-IJCNLP 2021 paper entitled "T-DNA: Taming Pre-trained Language Models with N-gram Representations for Low-Resource Domain Adaptation" by Shizhe Diao et al.

    Python 19 4

  6. extreme-bert/extreme-bert extreme-bert/extreme-bert Public

    ExtremeBERT is a toolkit that accelerates the pretraining of customized language models on customized datasets, described in the paper “ExtremeBERT: A Toolkit for Accelerating Pretraining of Custom…

    Python 269 14