AMDAI Compiler Development Engineer
任职要求
Job Summary: We are seeking a highly skilled Senior Member of Technical Staff (SMTS) Compiler Software Engineer to join our team. In this role, you will be responsible for designing, developing, and optimizing compiler solutions that drive the performance and efficiency of our cutting-edge hardware platforms. You will collaborate with a dynamic team of engineers to push the boundaries of compiler technology, enabling innovative solutions for complex software challenges. Key Responsibilities: Design and implement advanced compiler optimizations and code generation techniques to maximize performance for AMD NPUs. Develop and maintain core components of the compiler, including front-end, middle-end, and back-end. Analyze and optimize code performance, ensuring alignment with target platform requirements. Collaborate closely with hardware and software teams to define and implement features that enhance overall system performance. Conduct performance analysis, debug issues, and provide solutions for compiler-related problems. Stay up-to-date with the latest advancements in compiler technology, programming languages, and hardware trends. Provide technical guidance and mentorship to junior engineers. Qualifications: Education: Master’s, or Ph.D. in Computer Science, Computer Engineering, or a related field. Experience: Minimum of 5+ years of experience in compiler development or a related field. Technical Skills: Proficiency in C/C++ and familiarity with modern programming practices. Deep understanding of compiler architecture, optimization techniques, and code generation. Experience with LLVM, TVM, MLIR, or other compiler toolchains. Strong background in performance tuning and debugging. Familiarity with hardware architectures such as GPUs, CPUs, or custom accelerators. Soft Skills: Excellent problem-solving skills, strong analytical thinking, and effective communication skills. Preferred Qualifications: Experience with domain-specific languages (DSLs) or designing custom compiler extensions. Familiarity with machine learning frameworks and their optimization for hardware acceleration. Knowledge of scripting languages such as Python for automation tasks. Prior contributions to open-source compiler projects. #LI-FL1
工作职责
N/A
THE ROLE: MTS Software development engineer on teams building and optimizing Deep Learning applications and AI frameworks for AMD GPU compute platforms. Work as part of an AMD development team and open-source community to analyze, develop, test and deploy improvements to make AMD the best platform for machine learning applications. THE PERSON: Strong technical and analytical skills in C++ development in a Linux environment. Ability to work as part of a team, while also being able to work independently, define goals and scope and lead your own development effort. KEY RESPONSIBILITIES: Optimize Deep Learning Frameworks: In depth experience in enhance and optimize frameworks like TensorFlow and PyTorch for AMD GPUs in open-source repositories. Develop GPU Kernels: Create and optimize GPU kernels to maximize performance for specific AI operations. Develop & Optimize Models: Design and optimize deep learning models specifically for AMD GPU performance. Collaborate with GPU Library Teams: Work tightly with internal teams to analyze and improve training and inference performance on AMD GPUs. Collaborate with Open-Source Maintainers: Engage with framework maintainers to ensure code changes are aligned with requirements and integrated upstream. Work in Distributed Computing Environments: Optimize deep learning performance on both scale-up (multi-GPU) and scale-out (multi-node) systems. Utilize Cutting-Edge Compiler Tech: Leverage advanced compiler technologies to improve deep learning performance. Optimize Deep Learning Pipeline: Enhance the full pipeline, including integrating graph compilers. Software Engineering Best Practices: Apply sound engineering principles to ensure robust, maintainable solutions.
THE ROLE: Triton is a language and compiler for writing highly efficient custom deep learning primitives. It's widely adopted in open AI software stack projects like PyTorch, vLLM, SGLang, and many others. AMD GPU is an official backend in Triton and we are fully committed to it. If you are interested in making GPUs running fast via developing the Triton compiler and kernels, please come join us!
An exciting internship opportunity to make an immediate contribution to AMD's next generation of technology innovations awaits you! We have a multifaceted, high-energy work environment filled with a diverse group of employees, and we provide outstanding opportunities for developing your career. During your internship, our programs provide the opportunity to collaborate with AMD leaders, receive one-on-one mentorship, attend amazing networking events, and much more. Being part of AMD means receiving hands-on experience that will give you a competitive edge. Together We Advance your career! JOB DETAILS: Location: Beijing,China Onsite/Hybrid: at least 3 days a week, either in a hybrid or onsite or remote work structure throughout the duration of the co-op/intern term. Duration: at least 6 months WHAT YOU WILL BE DOING: We are seeking highly motivated AI Compiler Software Engineering intern/co-op to join our team. In this role – We will involve you in extending Triton’s compiler infrastructure to support new AI workloads and hardware targets. We will assign you tasks to implement and optimize GPU kernels using Triton’s Python-based DSL. We will train you to analyze kernel performance using profiling tools and help you identify bottlenecks and optimization opportunities. We will understand how modern compilers translate high-level abstractions into efficient machine code.
1、参与人工智能芯片的软硬件协同设计,指令集功能验证; 2、参与人工智能芯片的编译器算法设计和实现, 工具链开发与维护,网络模型的性能调优; 3、参与深度学习软件栈的设计和实现; 1. Working closely with hardware/architecture engineering and software teams to understand the hardware and software requirements. 2. Responsible for compiler and tool chain design, implementation, maintaining and performance tuning. 3. Responsible for the design and implementation of deep learning software stack.