New job, posted less than a week ago!
Job Details
Posted date: Aug 14, 2025
There have been 1068 jobs posted with the title of Principal Software Engineer all time at Microsoft.There have been 1068 Principal Software Engineer jobs posted in the last month.
Category: Software Engineering
Location: Redmond, WA
Estimated salary: $222,050
Range: $139,900 - $304,200
Employment type: Full-Time
Travel amount: 25.0%
Work location type: Up to 50% work from home
Role: Individual Contributor
Description
We are the AI Frameworks team at Microsoft, part of the CoreAI organization, and we are at the forefront of driving innovation in large-scale AI. Our mission is to enable state-of-the-art large language model (LLM) training and inference through deep optimization across the full software and hardware stack. We work across PyTorch, ONNX, and other open AI frameworks, pushing the boundaries of performance, scalability, and efficiency on a wide range of hardware accelerators.This position is part of a highly specialized sub-team focused on building the end-to-end software stack for Microsoft’s first-party (1P) AI accelerators. Our work is foundational to Microsoft’s long-term AI infrastructure strategy, powering next-generation models and services across Azure and Microsoft products. We collaborate closely with hardware architects, compiler teams, and model experts to co-design software solutions that unlock the full potential of custom silicon. The work is deeply technical and impactful—advancing AI systems used at global scale.
If you are passionate about AI systems, low-level performance optimization, and solving hard cross-discipline engineering problems, we invite you to join us and help shape the future of AI at Microsoft.
Design, implement, and optimize core components of the AI software stack targeting Microsoft’s first-party AI accelerators, including runtime, kernel libraries, and framework integration layers. Collaborate with hardware, compiler, and model teams to co-design solutions that maximize performance, efficiency, and reliability across the full AI stack.Develop performance-critical infrastructure to support inference of large language models (LLMs) at scale. Identify and address software bottlenecks, and drive end-to-end performance tuning and debugging across framework, runtime, and hardware layers.Work closely with partner teams across Azure, research, and product groups to align technical direction and deliver high-impact capabilities for real-world AI workloads.Participate in design reviews, code reviews, and architectural discussions to ensure high-quality and maintainable software.Stay current with advancements in AI frameworks, compiler technologies, and hardware acceleration, and bring relevant innovations into our software stack.
Qualifications
Required Qualifications:Bachelor's Degree in Computer Science or related technical field AND 6+ years technical engineering experience with coding in languages including, but not limited to, C, C++, or Python OR equivalent experience.Experience in C++ and/or Python, with deep understanding of software design, debugging, and performance optimization.Hands on experience with modern large language model (LLM) serving technologies, including model partitioning, distributed execution, and inference optimization.Experienced designing and delivering complex, high-performance systems in production environments.
Preferred Qualifications:
Master's Degree in Computer Science AND 8+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, or Python.Experience with deep learning frameworks such as PyTorch, TensorFlow, or ONNX.Experience with AI accelerator software stacks, including custom runtimes, graph compilers, kernel libraries, or device drivers.Experience with GPU computing, including CUDA programming, GPU kernel optimization, and performance tuning for large-scale AI workloads.Excellent cross-discipline collaboration skills; ability to work effectively with hardware, compiler, and ML model teams.Deep understanding of large-scale distributed training or inference systems for LLMs.Familiarity with low-level performance tools, profiling, and memory/compute efficiency analysis.Prior contributions to open-source AI frameworks or compilers (e.g., PyTorch, XLA, TVM, Triton).Technical leadership and mentorship experience; ability to lead by influence and drive cross-team alignment.Software Engineering IC5 - The typical base pay range for this role across the U.S. is USD $139,900 - $274,800 per year. There is a different range applicable to specific work locations, within the San Francisco Bay area and New York City metropolitan area, and the base pay range for this role in those locations is USD $188,000 - $304,200 per year.
Certain roles may be eligible for benefits and other compensation. Find additional benefits and pay information here: https://careers.microsoft.com/us/en/us-corporate-pay
Microsoft will accept applications and processes offers for these roles on an ongoing basis.
Microsoft is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to age, ancestry, color, family or medical care leave, gender identity or expression, genetic information, marital status, medical condition, national origin, physical or mental disability, political affiliation, protected veteran status, race, religion, sex (including pregnancy), sexual orientation, or any other characteristic protected by applicable laws, regulations and ordinances. We also consider qualified applicants regardless of criminal histories, consistent with legal requirements. If you need assistance and/or a reasonable accommodation due to a disability during the application or the recruiting process, please send a request via the Accommodation request form.
Benefits/perks listed below may vary depending on the nature of your employment with Microsoft and the country where you work.
#coreai
Check out other jobs at Microsoft.