About the job
Job Description\n\nNVIDIA has been transforming computer graphics, PC gaming, and accelerated computing for more than 25 years. It’s a unique legacy of innovation that’s fueled by great technology—and amazing people. Today, we’re tapping into the unlimited potential of AI to define the next era of computing.\n\nAs a new graduate, you'll help build the agentic infrastructure powering test automation and quality workflows for the NVIDIA Omniverse platform. This is a rare chance to initiate your career at the intersection of AI agents and production software quality. You will learn to build tests and tools other engineers depend on to ship quickly and confidently.\n\nWhat you'll be doing: Build multi-agent pipelines for automated test generation, log analysis, failure triage, and bug-filing workflows, working alongside senior engineers on well-scoped pieces of the system Contribute to evaluation systems that measure agent output quality — writing test cases, analyzing failure patterns, and extending eval frameworks under senior mentorship Add instrumentation, logging, and monitoring to agentic workflows so failures are visible and debuggable — learning the systems-thinking that makes infrastructure trustworthy Grow your judgment on where LLMs help and where they fail. Learn how to build solutions around both with mentorship.\n\nWhat we need to see: Pursuing or recently completed a Bachelor's Degree in Computer Science or equivalent Strong Python fundamentals — able to write clean, testable code and reason about structure beyond single scripts Hands-on exposure to AI-native development workflows — Claude Code, Cursor, Codex, or prompt engineering through coursework, internships, hackathons, or personal projects At least one project, open-source contribution, or coursework example where you coordinated an LLM into a working system end-to-end Foundational understanding of software testing, CI/CD concepts, or quality engineering principles Awareness of common LLM failure modes — hallucination, context limits, tool misuse — and curiosity about how to mitigate them\n\nWays to stand out from the crowd: Built a side project, hackathon entry, or open-source contribution involving multi-agent systems, MCP servers, or custom LLM tool integrations that you can walk through end-to-end Experimented with evaluating LLM outputs — even a small eval harness or scoring script for a personal project demonstrates the right instinct You have shipped something others actually used. It could be a tool, script, or bot adopted by a club, lab, or open-source community. You also provided documentation that let people use it without you You show intellectual integrity about where your projects break and have built in recovery paths rather than hiding failures