Job Description
What To Expect
Consider before submitting an application: This position is expected to start around January 2025 and continue through the entire Winter term (i.e. through May 2025) or into Summer 2025 if available. We ask for a minimum of 12 weeks, full-time and on-site, for most internships.
International Students: If your work authorization is through CPT, please consult your school on your ability to work 40 hours per week before applying. You must be able to work 40 hours per week on-site. Many students will be limited to part-time during the academic year.
About The Team
In this role, you will be responsible for the internal working of the AI inference stack and compiler running neural networks in millions of Tesla vehicles and Optimus. You will collaborate closely with the AI Engineers and Hardware Engineers to understand the full inference stack and design the compiler to extract the maximum performance out of our hardware.
The inference stack development is purpose-driven: deployment and analysis of production models inform the team's direction, and the team's work immediately impacts performance and the ability to deploy more and more complex models. With a cutting-edge co-designed MLIR compiler and runtime architecture, and full control of the hardware, the compiler has access to traditionally unavailable features, that can be leveraged via novel compilation approaches to generate higher performance models.
What You'll Do Take ownership of parts of AI Inference stack (Export/Compiler/Runtime) (flexible, based on skills/interests/needs)Closely collaborate with AI team to guide them on the design and the development of Neural Networks into production Collaborate with HW team to understand current HW architecture and propose future improvementsDevelop algorithms to improve performance and reduce compiler overheadDebug functional and performance issues on massively-parallel systemsWork on architecture-specific neural network optimization algorithms for high performance computing
What You'll Bring Pursuing a degree in Computer Science, Computer Engineering, or relevant field of study with a graduation date between 2025 -2026Must be able to relocate and work on site in Palo Alto, CAStrong C++ programming skills and familiarity with PythonSolid understanding of machine learning concepts and fundamentalsCapable of delivering results with minimal oversightExperience with quantization, MLIR, CUDA, and LLMs is a huge plus
Benefits
Compensation and Benefits
As a full-time Tesla Intern, you will be eligible for:
Aetna PPO and HSA plans > 2 medical plan options with $0 payroll deduction Family-building, fertility, adoption and surrogacy benefits Dental (including orthodontic coverage) and vision plans. Both have an option with a $0 payroll contribution Company Paid (Health Savings Account) HSA Contribution when enrolled in the High Deductible Medical Plan with HSA Healthcare and Dependent Care Flexible Spending Accounts (FSA) LGBTQ+ care concierge services 401(k), Employee Stock Purchase Plans, and other financial benefits Company Paid Basic Life, AD&D, and short-term disability insurance Employee Assistance Program Sick time after 90 days of employment and Paid Holidays Back-up childcare and parenting support resources Voluntary benefits to include: critical illness, hospital indemnity, accident insurance, theft & legal services, and pet insurance Commuter benefits Employee discounts and perks program
Expected Compensation
$100,000.0 - $150,000.0/annual salary + benefits
Pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. The total compensation package for this position may also include other elements dependent on the position offered. Details of participation in these benefit plans will be provided if an employee receives an offer of employment.
, Tesla
Job Tags
Holiday work, Full time, Temporary work, Part time, Summer work, Immediate start, Relocation, Flexible hours,