微软Principal Architect (Copilot Mac Team)
任职要求
Required Qualifications: 15+ years of software development experience10+ years of mobile development experience (iOS focus)5+ years of experience with C++, Objective-C Preferred Qualifications: Cross-platform development experience (Windows, Linux, macOS) is a plusExperience with web application development is a plusProven productivity and effectiveness in large codebases and/or open-source projects Familiarity with agile development practices, including daily standups; experience as a Scrum Master is a plus Microsoft is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to age, ancestry, color, family o…
工作职责
As an architect on the Copilot Mac team, you will drive the ongoing development of the product on macOS. The team is committed to delivering a consistent cross-platform experience, using data-driven methods to measure impact, reach, and reliability, while working closely with various internal teams. You will be responsible for the design, implementation, measurement, rollout, and refinement of solutions.
Architect, build, and optimize secure and performant AI platform services that power Microsoft Copilot and other next-generation AI scenarios. Provide technical leadership across teams to define long-term architectural direction and drive engineering excellence. Collaborate with infrastructure, platform, product, and research teams to design and deliver scalable, production-grade AI services. Write high-quality, well-tested, secure, and maintainable code and promote high standards across the team. Tackle technically ambiguous or cross-boundary problems, remove roadblocks, and drive delivery across multiple teams or organizations. Lead technical design discussions, mentor senior engineers, and foster a strong engineering culture within the team. Embody Microsoft’s Culture and Values, and help shape the direction of the engineering team and broader organization.
Architect, build, and optimize secure and performant AI platform services that power Microsoft Copilot and other next-generation AI scenarios. Provide technical leadership across teams to define long-term architectural direction and drive engineering excellence. Collaborate with infrastructure, platform, product, and research teams to design and deliver scalable, production-grade AI services. Write high-quality, well-tested, secure, and maintainable code and promote high standards across the team. Tackle technically ambiguous or cross-boundary problems, remove roadblocks, and drive delivery across multiple teams or organizations. Lead technical design discussions, mentor senior engineers, and foster a strong engineering culture within the team. Embody Microsoft’s Culture and Values, and help shape the direction of the engineering team and broader organization.
Responsibilities Collaborate with GPU sales team and SCE AIML TPM team to provide technical support for customers both at pre-sales and after-sales stage. Take ownership of problems and work to identify solutions. Design, deploy, and manage infrastructure components such as cloud resources, distributed computing systems, and data storage solutions to support AI/ML workflows. Collaborate with customers’ scientists and software/infrastructure engineers to understand infrastructure requirements for training, testing, and deploying machine learning models. Implement automation solutions for provisioning, configuring, and monitoring AI/ML infrastructure to streamline operations and enhance productivity. Optimize infrastructure performance by tuning parameters, optimizing resource utilization, and implementing caching and data pre-processing techniques. Troubleshoot infrastructure performance, scalability, and reliability issues and implement solutions to mitigate risks and minimize downtime. Stay updated on emerging technologies and best practices in AI/ML infrastructure and evaluate their potential impact on our systems and workflows. Document infrastructure designs, configurations, and procedures to facilitate knowledge sharing and ensure maintainability. Qualifications: Experience in scripting and automation using tools like Ansible, Terraform, and/or Kubernetes. Experience with containerization technologies (e.g., Docker, Kubernetes) and orchestration tools for managing distributed systems. Solid understanding of networking concepts, security principles, and best practices. Excellent problem-solving skills, with the ability to troubleshoot complex issues and drive resolution in a fast-paced environment. Strong communication and collaboration skills, with the ability to work effectively in cross-functional teams and convey technical concepts to non-technical stakeholders. Strong documentation skills with experience documenting infrastructure designs, configurations, procedures, and troubleshooting steps to facilitate knowledge sharing, ensure maintainability, and enhance team collaboration. Strong Linux skills with hands-on experience in Oracle Linux/RHEL/CentOS, Ubuntu, and Debian distributions, including system administration, package management, shell scripting, and performance optimization.
- As an AIML Specialist Solutions Architect (SA) in AI Infrastructure, you will serve as the Subject Matter Expert (SME) for providing optimal solutions in model training and inference workloads that leverage Amazon Web Services accelerator computing services. As part of the Specialist Solutions Architecture team, you will work closely with other Specialist SAs to enable large-scale customer model workloads and drive the adoption of AWS EC2, EKS, ECS, SageMaker and other computing platform for GenAI practice. - You will interact with other SAs in the field, providing guidance on their customer engagements, and you will develop white papers, blogs, reference implementations, and presentations to enable customers and partners to fully leverage AI Infrastructure on Amazon Web Services. You will also create field enablement materials for the broader SA population, to help them understand how to integrate Amazon Web Services GenAI solutions into customer architectures. - You must have deep technical experience working with technologies related to Large Language Model (LLM), Stable Diffusion and many other SOTA model architectures, from model designing, fine-tuning, distributed training to inference acceleration. A strong developing machine learning background is preferred, in addition to experience building application and architecture design. You will be familiar with the ecosystem of Nvidia and related technical options, and will leverage this knowledge to help Amazon Web Services customers in their selection process. - Candidates must have great communication skills and be very technical and hands-on, with the ability to impress Amazon Web Services customers at any level, from ML engineers to executives. Previous experience with Amazon Web Services is desired but not required, provided you have experience building large scale solutions. You will get the opportunity to work directly with senior engineers at customers, partners and Amazon Web Services service teams, influencing their roadmaps and driving innovations.