-
for performance, cost-efficiency, and low-latency inference Develop distributed model training and inference architectures leveraging GPU-based compute resources Implement server-less and containerized solutions
-
conferences. Qualifications: PhD in computer science with file systems, GPU architecture experience. Proven ability to articulate research work and findings in peer-reviewed proceedings. Knowledge of systems
Enter an email to receive alerts for gpu "https:" positions