Job Description

We are now looking for an AI Computing Software Development Intern!


NVIDIA invites skilled interns in artificial intelligence computing solutions to join our AI Compute team in Taiwan. This is your chance to work on one of the globe’s most advanced AI systems. You will help develop technologies for Large Language Models, Recommender Systems, and Generative AI, and push the limits of GPU performance for AI inference.

What you'll be doing:
As an intern, you’ll focus on one of two specialized tracks: TensorRT-LLM – Inference Optimization (Python / PyTorch) or TensorRT Compiler – Graph Optimization (C++).

For TensorRT-LLM:
+ Build and enhance high‑performance LLM inference pipelines.
+ Analyze and optimize model execution, scalability, and memory use.
+ Collaborate across framework and research teams to deliver efficient multi‑GPU model serving.


For TensorRT Compiler:
+ Work on the TensorRT compiler backend to improve graph tra...

Apply for this Position

Ready to join NVIDIA? Click the button below to submit your application.

Submit Application