mirror of https://github.com/hpcaitech/ColossalAI
[shardformer] fix embedding
parent
922302263b
commit
73a4144b91
|
@ -214,6 +214,9 @@ class VocabParallelEmbedding1D(ParallelModule):
|
||||||
self.vocab_start_index = tensor_parallel_rank * self.num_embeddings_per_partition
|
self.vocab_start_index = tensor_parallel_rank * self.num_embeddings_per_partition
|
||||||
self.vocab_end_index = self.vocab_start_index + self.num_embeddings_per_partition
|
self.vocab_end_index = self.vocab_start_index + self.num_embeddings_per_partition
|
||||||
|
|
||||||
|
# padding index
|
||||||
|
self.padding_idx = self._select_padding_idx(padding_idx)
|
||||||
|
|
||||||
# offset the seed with randomizer index and rank
|
# offset the seed with randomizer index and rank
|
||||||
seed = torch.random.initial_seed()
|
seed = torch.random.initial_seed()
|
||||||
self.randomizer = create_randomizer_with_offset(seed, process_group=self.process_group)
|
self.randomizer = create_randomizer_with_offset(seed, process_group=self.process_group)
|
||||||
|
|
Loading…
Reference in New Issue