GRPO/DPO Training Job
Production configuration for running NeMo-RL (GRPO) and TRL (DPO) training jobs with labeled preference datasets on Karpenter Spot node pools and Volcano Gang Scheduling.
Production configuration for running NeMo-RL (GRPO) and TRL (DPO) training jobs with labeled preference datasets on Karpenter Spot node pools and Volcano Gang Scheduling.
NVIDIA NeMo Framework distributed training, fine-tuning, and TensorRT-LLM conversion architecture
A hybrid ML architecture that trains on SageMaker and serves on EKS