As a Software Engineer within our Autonomy teams, you will contribute to one of the most advanced and widely deployed AI Platforms in the world for Autopilot and our Humanoid Robot, Optimus.
In this role, you will be responsible for the internal working of the AI inference stack and runtime running neural networks in millions of Tesla vehicles and Optimus, collaborating closely with the AI Engineers and Hardware Engineers. The nature of the role is multi-disciplinary - you will be working at the intersection of machine learning and systems. You will build the frameworks and infrastructure that enable the seamless deployment, integration, and inference of all neural networks. You will develop system tools to benchmark, characterize, and optimize the latency and throughput of AI workloads on the FSD chip.
The inference stack development is purpose-driven: deployment and analysis of production models informs the team's direction, and the team's work immediately impacts performance and the ability to deploy more and more complex models. With a cutting-edge co-designed runtime and MLIR compiler architecture, and full control of the hardware, the runtime enables extremely efficient use of features traditionally hidden behind layers of abstractions.