Skip to content

Commit

Permalink
update config hub
Browse files Browse the repository at this point in the history
  • Loading branch information
rasbt committed May 22, 2024
1 parent 379e54e commit eb6ec38
Show file tree
Hide file tree
Showing 29 changed files with 80 additions and 36 deletions.
1 change: 1 addition & 0 deletions config_hub/finetune/falcon-7b/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -113,6 +113,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/falcon-7b/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -115,6 +115,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/gemma-2b/full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -84,6 +84,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/gemma-2b/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/gemma-2b/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/gemma-7b/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/gemma-7b/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/llama-2-7b/full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -87,6 +87,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/llama-2-7b/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -113,6 +113,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/llama-2-7b/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -115,6 +115,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/llama-3-8b/full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -87,6 +87,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/llama-3-8b/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -113,6 +113,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/llama-3-8b/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -115,6 +115,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/mistral-7b-v0.2/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -113,6 +113,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/mistral-7b-v0.2/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -115,6 +115,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/mistral-7b/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -113,6 +113,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/mistral-7b/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -115,6 +115,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/phi-2/full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -87,6 +87,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/phi-2/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/phi-2/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/stablelm-base-alpha-3b/full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -84,6 +84,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/stablelm-base-alpha-3b/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -113,6 +113,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/stablelm-base-alpha-3b/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -115,6 +115,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/tiny-llama/full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -84,6 +84,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/tiny-llama/lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
1 change: 1 addition & 0 deletions config_hub/finetune/tiny-llama/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,7 @@ seed: 1337

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:
Expand Down
30 changes: 18 additions & 12 deletions config_hub/pretrain/debug.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -58,18 +58,6 @@ train:
# Whether to tie the embedding weights with the language modeling head weights. (type: Optional[bool], default: False)
tie_embeddings:

# (type: float, default: 0.0004)
learning_rate: 6e-4

# (type: float, default: 0.1)
weight_decay: 0.1

# (type: float, default: 0.9)
beta1: 0.9

# (type: float, default: 0.95)
beta2: 0.95

# (type: Optional[float], default: 1.0)
max_norm: 1.0

Expand All @@ -91,6 +79,24 @@ eval:
# Whether to evaluate on the validation set at the beginning of the training
initial_validation: false

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:

# (type: float, default: 0.001)
lr: 6e-4

# (type: float, default: 0.01)
weight_decay: 0.1

# (type: tuple, default: (0.9,0.999))
betas:
- 0.9
- 0.95

# How many devices/GPUs to use. Uses all GPUs by default. (type: Union[int, str], default: auto)
devices: auto

Expand Down
30 changes: 18 additions & 12 deletions config_hub/pretrain/tinyllama.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -58,18 +58,6 @@ train:
# Whether to tie the embedding weights with the language modeling head weights. (type: Optional[bool], default: False)
tie_embeddings:

# (type: float, default: 0.0004)
learning_rate: 4e-4

# (type: float, default: 0.1)
weight_decay: 0.1

# (type: float, default: 0.9)
beta1: 0.9

# (type: float, default: 0.95)
beta2: 0.95

# (type: Optional[float], default: 1.0)
max_norm: 1.0

Expand All @@ -91,6 +79,24 @@ eval:
# Whether to evaluate on the validation set at the beginning of the training
initial_validation: false

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:

# (type: float, default: 0.001)
lr: 4e-4

# (type: float, default: 0.01)
weight_decay: 0.1

# (type: tuple, default: (0.9,0.999))
betas:
- 0.9
- 0.95

# How many devices/GPUs to use. Uses all GPUs by default. (type: Union[int, str], default: auto)
devices: auto

Expand Down
30 changes: 18 additions & 12 deletions config_hub/pretrain/tinystories.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -74,18 +74,6 @@ train:
# Whether to tie the embedding weights with the language modeling head weights. (type: Optional[bool], default: False)
tie_embeddings: true

# (type: float, default: 0.0004)
learning_rate: 0.0005

# (type: float, default: 0.1)
weight_decay: 0.1

# (type: float, default: 0.9)
beta1: 0.9

# (type: float, default: 0.95)
beta2: 0.95

# (type: Optional[float], default: 1.0)
max_norm: 1.0

Expand All @@ -107,6 +95,24 @@ eval:
# Whether to evaluate on the validation set at the beginning of the training
initial_validation: false

# Optimizer-related arguments
optimizer:

class_path: torch.optim.AdamW

init_args:

# (type: float, default: 0.001)
lr: 0.0005

# (type: float, default: 0.01)
weight_decay: 0.1

# (type: tuple, default: (0.9,0.999))
betas:
- 0.9
- 0.95

# How many devices/GPUs to use. Uses all GPUs by default. (type: Union[int, str], default: auto)
devices: auto

Expand Down

0 comments on commit eb6ec38

Please sign in to comment.