Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEAT] LLM을 QLoRA 방식으로 fine-tuning하여 페르소나 추출 모델 만들기 #32

Closed
3 tasks done
DonghaeSuh opened this issue Mar 29, 2024 · 0 comments · Fixed by #33
Closed
3 tasks done
Assignees

Comments

@DonghaeSuh
Copy link
Contributor

DonghaeSuh commented Mar 29, 2024

Category

  • Model

Background

  • QLoRA라는 Quantization이후, LoRA방식으로 LLM 모델 fine-tuning

To do

  • LLM 모델 선정
  • Peft 라이브러리를 활용한 QLoRA 학습
  • Inference 결과 확인

Expected behavior

  • LLM에 정제된 적은 데이터(2,330)개를 fine-tuning한 경우 모델이 사용자 발화에서 페르소나를 잘 추출해내는지 확인 가능
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant