Heegyu Kim PRO
heegyu
AI & ML interests
NLP
Recent Activity
View all activity
Organizations
Korean Reward Modeling
Korean Datasets, Reward Models for RLHF
-
heegyu/KoSafeGuard-8b-0503
Text Generation • 8B • Updated • 43 • 5 -
heegyu/ko-reward-model-helpful-1.3b-v0.2
Text Classification • 1B • Updated • 10 -
heegyu/ko-reward-model-safety-1.3b-v0.2
Text Classification • 1B • Updated • 12 • 5 -
heegyu/ko-reward-model-helpful-roberta-large-v0.1
Text Classification • 0.3B • Updated • 13 • 1
AjouBlue GPTs
AjouBlue-GPT: Small Korean Generative Language Models with 125M, 355M parameters
Synthetic Dataset
Reward Modeling Datasets
Vision LM
Domain Specific (Math, Code, etc)
Safety LM
R1-like Datasets
Korean Pretraining Dataset
Datasets Translated to Korean
RLHF papers
-
Pairwise Proximal Policy Optimization: Harnessing Relative Feedback for LLM Alignment
Paper • 2310.00212 • Published • 2 -
Stabilizing RLHF through Advantage Model and Selective Rehearsal
Paper • 2309.10202 • Published • 11 -
Aligning Language Models with Offline Reinforcement Learning from Human Feedback
Paper • 2308.12050 • Published • 1 -
Secrets of RLHF in Large Language Models Part I: PPO
Paper • 2307.04964 • Published • 29
Pre-training Dataset
Image Generation
Machine Translation
Text2SQL
SWE Agent
R1-like Datasets
Korean Reward Modeling
Korean Datasets, Reward Models for RLHF
-
heegyu/KoSafeGuard-8b-0503
Text Generation • 8B • Updated • 43 • 5 -
heegyu/ko-reward-model-helpful-1.3b-v0.2
Text Classification • 1B • Updated • 10 -
heegyu/ko-reward-model-safety-1.3b-v0.2
Text Classification • 1B • Updated • 12 • 5 -
heegyu/ko-reward-model-helpful-roberta-large-v0.1
Text Classification • 0.3B • Updated • 13 • 1
Korean Pretraining Dataset
AjouBlue GPTs
AjouBlue-GPT: Small Korean Generative Language Models with 125M, 355M parameters
Datasets Translated to Korean
Synthetic Dataset
RLHF papers
-
Pairwise Proximal Policy Optimization: Harnessing Relative Feedback for LLM Alignment
Paper • 2310.00212 • Published • 2 -
Stabilizing RLHF through Advantage Model and Selective Rehearsal
Paper • 2309.10202 • Published • 11 -
Aligning Language Models with Offline Reinforcement Learning from Human Feedback
Paper • 2308.12050 • Published • 1 -
Secrets of RLHF in Large Language Models Part I: PPO
Paper • 2307.04964 • Published • 29
Reward Modeling Datasets
Pre-training Dataset
Vision LM
Image Generation
Domain Specific (Math, Code, etc)
Machine Translation
Safety LM
Text2SQL