Skip to content

Commit 9fb12d5

Browse files
committed
.
1 parent ae9f81f commit 9fb12d5

File tree

3 files changed

+4
-3
lines changed

3 files changed

+4
-3
lines changed

configs/125M.yml

+1
Original file line numberDiff line numberDiff line change
@@ -51,6 +51,7 @@
5151
# batch / data settings
5252
"train_micro_batch_size_per_gpu": 4,
5353
"data-impl": "mmap",
54+
"gradient_accumulation_steps": 1,
5455

5556
# activation checkpointing
5657
"checkpoint-activations": true,

configs/local_setup.yml

+2-1
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,6 @@
11
# Suggested data paths when using GPT-NeoX locally
22
{
3+
"global_num_gpus":1,
34
"data-path": "data/enwik8/enwik8_text_document",
45

56
# or for weighted datasets:
@@ -24,7 +25,7 @@
2425

2526
"tensorboard-dir": "tensorboard",
2627
"log-dir": "logs",
27-
"use_wandb": True,
28+
"use_wandb": False,
2829
"wandb_host": "https://api.wandb.ai",
2930
"wandb_project": "neox"
3031
}

megatron/neox_arguments/neox_args.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -941,8 +941,7 @@ class NeoXArgsTraining(NeoXArgsTemplate):
941941
"""
942942

943943
gas: int = None
944-
"""gradient_accumulation_steps""" # TODO this is a duplicate, remove?
945-
944+
gradient_accumulation_steps: int = 1
946945
clip_grad: float = None
947946
"""
948947
Gradient clipping based on global L2 norm.

0 commit comments

Comments
 (0)