Skip to content

Commit

Permalink
Add Support for GPT-2 Training on different Devices (#551)
Browse files Browse the repository at this point in the history
* fix bfloat16

* update tokenizer encode

* support python 3.10-

* support python 3.10-

* support python 3.10-

* click

* fix tokenizer path

* update

* update

* checkout main files
  • Loading branch information
ShawnXuan authored Sep 13, 2024
1 parent 7aeb4ca commit 13056f4
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 2 deletions.
4 changes: 3 additions & 1 deletion libai/models/gpt_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -244,7 +244,9 @@ def forward(self, input_ids, past_length=0):
bsz, seq_length = input_ids.size()

position_ids = self.position_ids[:, past_length : past_length + seq_length]
position_ids = position_ids.expand_as(input_ids).to_global(sbp=input_ids.sbp)
position_ids = position_ids.expand_as(input_ids).to_global(
sbp=input_ids.sbp, placement=input_ids.placement
)

token_embeds = self.token_embeddings(input_ids)
position_embeds = self.position_embeddings(position_ids)
Expand Down
3 changes: 2 additions & 1 deletion tools/train_net.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,8 @@ def main(args):

seed_for_rank = cfg.train.seed + flow.env.get_rank()
flow.manual_seed(seed_for_rank)
flow.cuda.manual_seed(seed_for_rank)
if flow.cuda.is_available():
flow.cuda.manual_seed(seed_for_rank)
np.random.seed(seed_for_rank)
random.seed(seed_for_rank)

Expand Down

0 comments on commit 13056f4

Please sign in to comment.