Sort by
Refine Your Search
-
optimization and LLM alignment: design preference-based training and fine-tuning methods (RLHF, PPO, DPO, reward modeling) for medical and multilingual LLMs. Agentic and tool-augmented AI systems: develop
Searches related to agent based modelling
Enter an email to receive alerts for agent-based-modelling positions