-
for Science @ Scale: Pretraining, instruction tuning, continued pretraining, Mixture-of-Experts; distributed training/inference (FSDP, DeepSpeed, Megatron-LM, tensor/sequence parallelism); scalable evaluation
Searches related to parallel and distributed computing phd
Enter an email to receive alerts for parallel-and-distributed-computing-phd positions