small fix on world_size with pgm

This commit is contained in:
ferdinand.mom 2024-12-02 18:12:02 +00:00
parent b6267c768e
commit 75939867d9

View File

@ -126,10 +126,10 @@ if __name__ == "__main__":
project="picotron", project="picotron",
name=f"{config['logging']['run_name']}_{tokens_per_step}_{pgm.process_group_manager}", name=f"{config['logging']['run_name']}_{tokens_per_step}_{pgm.process_group_manager}",
config={ config={
"tensor_parallel_size": pgm.process_group_manager.tp_size, "tensor_parallel_size": pgm.process_group_manager.tp_world_size,
"context_parallel_size": pgm.process_group_manager.cp_size, "context_parallel_size": pgm.process_group_manager.cp_world_size,
"pipeline_parallel_size": pgm.process_group_manager.pp_size, "pipeline_parallel_size": pgm.process_group_manager.pp_world_size,
"data_parallel_size": pgm.process_group_manager.dp_size, "data_parallel_size": pgm.process_group_manager.dp_world_size,
"model": config["model"]["name"], "model": config["model"]["name"],
"dataset": config["dataset"]["name"], "dataset": config["dataset"]["name"],
"max_tokens": config["training"]["max_tokens"], "max_tokens": config["training"]["max_tokens"],