Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[事前学習] - Instruction pre-training実証実験 #95

Open
odashi opened this issue Dec 9, 2024 · 0 comments
Open

[事前学習] - Instruction pre-training実証実験 #95

odashi opened this issue Dec 9, 2024 · 0 comments
Assignees
Labels
pretrain Experiment of model pretrain

Comments

@odashi
Copy link
Member

odashi commented Dec 9, 2024

Overview

LLM-jpのbaseモデルに対してinstruction pre-trainingを行ったとき、どのように性能が変化するのかを検証する。

Details

1.8B, 3.7B, 7.2B, 13Bで実験を行い、172Bでの有望な設定を検証する。

使用するデータの候補:

  • Instruct2データセット

データの導入方法

  • LLM-jpコーパスに対して、特定の比率でinstructionデータを混入させる。
    • 混入量の候補: 10%, 1%, 0.1%

Resources

@odashi odashi added the pretrain Experiment of model pretrain label Dec 9, 2024
@odashi odashi changed the title [事前学習] - [事前学習] - Instruction pre-training実証実験 Dec 9, 2024
@odashi odashi self-assigned this Dec 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
pretrain Experiment of model pretrain
Projects
None yet
Development

No branches or pull requests

1 participant