LLM fine-tuning
with medical records 이론/실습
2025-10-23, 15:00 ~ 17:00
Seongsu Bae, Sujeong Im
KAIST AI @ Edlab (Advised by Edward Choi)
KoSAIM 2025 개발자를 위한 AI 실습교육: Train your own medical AI
Speaker Bio
Sujeong Im (임수정)
Education
Research Interests
Seongsu Bae (배성수)
Education
Research Interests
Table of Contents
Language Model
We deal with LMs every day!
How to train a LM?
The
The
sky
is
blue
.
The
sky
is
The
sky
blue
is
sky
Next Token Prediction task for the sentence “The sky is blue.”
Text Generation via a Probabilistic Model
The
sky
is
blue
clear
usually
the
<
(Large)
Language Model
more likely
less likely
How to build a (large) language model?
Pretrained
LM
Finetune on task A
Finetune on task B
Finetune on task C
Inference
on task A
Inference
on task B
Inference
on task C
(-) Task-specific training → One specialized model for each task
How to build a (large) language model?
Pretrained
LM
Inference
on task A
Inference
on task B
Inference
on task C
(+) Improve performance via few-shot prompting or prompt engineering
prompting
How to build a (large) language model?
(-) Forced few-shot prompting
(-) Manual efforts for the prompting technique
(-) Not aligned with natural instructions
How to build a (large) language model?
Pretrained
LM
Inference
on task A
Inference
on task B
Inference
on task C
(+) model learns to perform many tasks via natural language instructions
…
instructions
fine-tune on many instructions
How to build a (large) language model?
Building an instruction-following LLM
Imagine a clinical LLM
Asclepius: Publicly Shareable Clinical Large Language Model Built on Synthetic Clinical Notes (Gweon and Kim et al., ACL 2024 Findings)
Real clinical note
Case report
Synthetic clinical note generation
Clinical instruction/response data generation
Final dataset
Asclepius-Llama3-8B
Hands-on Session:
Fine-tuning a clinical domain LLM
Environment Setup
colab link
Environment Setup
Environment Setup
Colab Objectives
Deep learning memory layout
Can You Run it?
LoRA (Hu and Shen et al., 2021)
QLoRA (Dettmers and Pagnoni et al., 2023)
Parameter-Efficient Fine-Tuning (PEFT)
Thank you :D
If you require any further information, feel free to contact us: seongsu@kaist.ac.kr, sujeongim@kaist.ac.kr
KoSAIM 2025 개발자를 위한 AI 실습교육: Train your own medical AI