Skip to yearly menu bar Skip to main content


MLSys 2025 Career Opportunities

Here we highlight career opportunities submitted by our Exhibitors, and other top industry, academic, and non-profit leaders. We would like to thank each of our exhibitors for supporting MLSys 2025.

Search Opportunities

Santa Clara


Description:

AWS AI/ML is looking for world class scientists and engineers to join its AI Research and Education group working on foundation models, large-scale representation learning, and distributed learning methods and systems. At AWS AI/ML you will invent, implement, and deploy state of the art machine learning algorithms and systems. You will build prototypes and innovate on new representation learning solutions. You will interact closely with our customers and with the academic and research communities. You will be at the heart of a growing and exciting focus area for AWS and work with other acclaimed engineers and world famous scientists.

Large-scale foundation models have been the powerhouse in many of the recent advancements in computer vision, natural language processing, automatic speech recognition, recommendation systems, and time series modeling. Developing such models requires not only skillful modeling in individual modalities, but also understanding of how to synergistically combine them, and how to scale the modeling methods to learn with huge models and on large datasets. Join us to work as an integral part of a team that has diverse experiences in this space. We actively work on these areas:

  • Hardware-informed efficient model architecture, training objective and curriculum design
  • Distributed training, accelerated optimization methods
  • Continual learning, multi-task/meta learning
  • Reasoning, interactive learning, reinforcement learning
  • Robustness, privacy, model watermarking
  • Model compression, distillation, pruning, sparsification, quantization

Location: Redwood City, CA or New York, NY


About Us:

Here at Fireworks, we’re building the future of generative AI infrastructure. Fireworks offers the generative AI platform with the highest-quality models and the fastest, most scalable inference. We’ve been independently benchmarked to have the fastest LLM inference and have been getting great traction with innovative research projects, like our own function calling and multi-modal models. Fireworks is funded by top investors, like Benchmark and Sequoia, and we’re an ambitious, fun team composed primarily of veterans from Pytorch and Google Vertex AI.

The Role:

As a Technical Developer Advocate at Fireworks AI, you will serve as a technical ambassador for our state-of-the-art platform. In this role, you’ll focus on building strong relationships with developers, crafting compelling technical content, and sharing insights that directly influence our product evolution. This role calls for deep technical expertise, creative communication, and a genuine passion for community engagement.

Key Responsibilities:

  • Community Engagement: Foster a vibrant developer community through forums, social media, webinars, meetups, hackathons, and conferences. Drive initiatives that encourage connection, collaboration, and innovation.
  • Technical Evangelism: Deliver engaging presentations, live demos, and hands-on workshops that highlight Fireworks' capabilities and inspire platform adoption.
  • Content Creation & Thought Leadership: Produce high-quality technical content—blogs, tutorials, docs, and videos—that simplify complex AI topics and provide actionable value.
  • Advocacy & Feedback: Serve as a trusted advocate by capturing developer feedback and sharing insights with internal teams to inform product development.
  • Cross-Functional Collaboration: Partner with Product and Engineering to translate developer needs into impactful roadmap decisions and innovations.
  • Impact Measurement: Leverage data and metrics to assess engagement effectiveness and optimize community strategies.

Minimum qualifications:

  • Bachelor’s degree in Computer Science, Engineering, or a related field—or equivalent practical experience.
  • 3+ years in developer relations, developer advocacy, technical evangelism, or similar roles within developer products or SaaS.
  • Proven experience in engaging and nurturing developer communities.
  • Strong technical proficiency in programming languages (e.g., Python, JavaScript) and familiarity with AI/ML concepts.
  • Excellent communication and presentation skills, with the ability to explain complex technical topics in an accessible manner.
  • A data-driven mindset with strong analytical abilities.

Preferred qualifications:

  • Experience in the AI or machine learning industry, with a deep understanding of generative AI technologies.
  • Previous experience in a startup or fast-paced tech environment.
  • Demonstrated success in creating compelling technical content for diverse developer audiences.
  • Experience contributing to and engaging with open-source communities.
  • Familiarity with tools for community engagement, content management, and performance analytics.

Why Fireworks AI?

  • Solve Hard Problems: Tackle challenges at the forefront of AI infrastructure, from low-latency inference to scalable model serving.
  • Build What’s Next: Work with bleeding-edge technology that impacts how businesses and developers harness AI globally.
  • Ownership & Impact: Join a fast-growing, passionate team where your work directly shapes the future of AI—no bureaucracy, just results.
  • Learn from the Best: Collaborate with world-class engineers and AI researchers who thrive on curiosity and innovation.

Fireworks AI is an equal-opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all innovators.

Location

Cupertino


Description

AWS Neuron is the complete software stack for the AWS Inferentia and Trainium cloud-scale machine learning accelerators. As a part of the Neuron Frameworks team you'll develop and enhance support PyTorch and JAX for AWS Neuron, working with the open source ecosystem.

You will develop and extend support for the leading ML frameworks, delivering an outstanding user experience for PyTorch and JAX ML model development on the Trainium and Inferentia accelerators. You will work closely with teams across AWS Neuron including compiler, training and inference optimization to optimize frameworks for AWS's accelerator architectures, and engage closely with the PyTorch and JAX and other ML Framework communities to take advantage of their latest capabilities and improve performance and usability for ML model developers.

A successful candidate will have a experience developing Machine Learning infrastructure and/or ML Frameworks, a demonstrated ability to work with open source communities to influence future community direction, a robust technical ability and a motivation to achieve results. Experience with technologies and tools such as XLA, vLLM or Hugging Face transformers is highly valued.

Utility Computing (UC) AWS Utility Computing (UC) provides product innovations — from foundational services such as Amazon’s Simple Storage Service (S3) and Amazon Elastic Compute Cloud (EC2), to consistently released new product innovations that continue to set AWS’s services and features apart in the industry. As a member of the UC organization, you’ll support the development and management of Compute, Database, Storage, Internet of Things (Iot), Platform, and Productivity Apps services in AWS, including support for customers who require specialized security solutions for their cloud services.

SF Bay Area or New York City


About the role We’re looking for seasoned ML Infrastructure engineers with experience designing, building and maintaining training and serving infrastructure for ML research.

Responsibilities:

Provide infrastructure support to our ML research and product

Build tooling to diagnose cluster issues and hardware failures

Monitor deployments, manage experiments, and generally support our research

Maximize GPU allocation and utilization for both serving and training

Requirements:

4+ years of experience supporting the infrastructure within an ML environment

Experience in developing tools used to diagnose ML infrastructure problems and failures

Experience with cloud platforms (e.g., Compute Engine, Kubernetes, Cloud Storage)

Experience working with GPUs

Nice to have

Experience with large GPU clusters and high-performance computing/networking

Experience with supporting large language model training

Experience with ML frameworks like Pytorch/TensorFlow/JAX

Experience with GPU kernel development

Location: Redwood City, CA or New York, NY


About Us:

Here at Fireworks, we’re building the future of generative AI infrastructure. Fireworks offers the generative AI platform with the highest-quality models and the fastest, most scalable inference. We’ve been independently benchmarked to have the fastest LLM inference and have been getting great traction with innovative research projects, like our own function calling and multi-modal models. Fireworks is funded by top investors, like Benchmark and Sequoia, and we’re an ambitious, fun team composed primarily of veterans from Pytorch and Google Vertex AI.

Job Overview

As an Applied Machine Learning Engineer, you will serve as a vital bridge between cutting-edge AI research and practical, real-world applications. Your work will focus on developing, fine-tuning, and operationalizing machine learning models that drive business value and enhance user experiences. This is a hands-on engineering role that combines deep technical expertise with a strong customer focus to deliver scalable AI solutions.

Responsibilities

  • Customer Success: Collaborate directly with the GTM team (Account Executives and Solutions Architects) to ensure smooth integration and successful deployment of ML solutions.
  • Demo / Proof of Concept (PoC): Build and present compelling PoCs that demonstrate the capabilities of our AI technology.
  • Application Build: Design, develop, and deploy end-to-end AI-powered applications tailored to customer needs.
  • Platform Features / Bug Fixes: Contribute to the internal ML platform, including adding features and resolving issues.
  • New Model Enablements: Integrate and enable new machine learning models into the existing platform or client environments.
  • Performance Optimizations: Improve system performance, efficiency, and scalability of deployed models and applications.
  • Partnership Enablement: Work closely with partners to enable joint AI solutions and ensure seamless collaboration.

Minimum Qualifications

  • Bachelor’s degree in Computer Science, Engineering, or a related technical field.
  • 5+ years of experience in a software engineering role, with a strong preference for customer-facing roles.
  • Robust coding skills required, preferably with proficiency in Python.
  • Demonstrated ability to lead and execute complex technical projects with a focus on customer success.
  • Strong interpersonal and communication skills; ability to thrive in dynamic, cross-functional teams.

Preferred Qualifications

  • Master’s degree in Computer Science, Engineering, or a related technical field.
  • Experience working in a startup or fast-paced environment.
  • Hands-on experience fine-tuning machine learning models, including supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF or RFT).
  • Solid understanding of generative AI, machine learning principles, and enterprise infrastructure.

Why Fireworks AI?

  • Solve Hard Problems: Tackle challenges at the forefront of AI infrastructure, from low-latency inference to scalable model serving.
  • Build What’s Next: Work with bleeding-edge technology that impacts how businesses and developers harness AI globally.
  • Ownership & Impact: Join a fast-growing, passionate team where your work directly shapes the future of AI—no bureaucracy, just results.
  • Learn from the Best: Collaborate with world-class engineers and AI researchers who thrive on curiosity and innovation.

Fireworks AI is an equal-opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all innovators.

San Francisco, California & Seattle, Washington

Are you a charismatic data engineering Developer Advocate at heart, someone who excels in the spotlight, loves presenting and hosting events, and possesses a deep knowledge of the Databricks Intelligence Platform?

As a Databricks Developer Advocate, you’ll be a crucial link between our engineering teams and the broad community of data professionals building their careers on Databricks technologies. This role will leverage your strong expertise in Databricks to educate, inspire, and support our growing user base.

Your responsibilities will encompass a wide range of knowledge-sharing activities. You’ll deliver engaging talks, host insightful panels and meetups, create informative blogs and video content, develop comprehensive courseware, provide expert answers in community forums, and conduct one-on-one sessions with influential data engineers and scientists. These efforts will help disseminate best practices and foster a thriving Databricks community.

Community engagement will be at the heart of your role. You’ll work closely with the Databricks community and the product and engineering teams, ensuring that user needs and product development align seamlessly. Reporting directly to the Head of Developer Relations, you’ll collaborate with fellow developer advocates and program managers to create and execute a cohesive and impactful developer relations strategy.

The ideal candidate for this position will embody the values of our Developer Relations team: a deep passion for data and AI, genuine empathy and technical understanding of developers’ needs, and a strong commitment to effectively explaining our products.

You’ll use your diverse Databricks skill set to educate the community about Databricks technologies and continuously gather and utilize community feedback to improve the developer experience.

The impact you will have:

- Create compelling content to inspire data practitioners, helping them discover new features and run projects at scale.
- Drive awareness and adoption of Databricks technologies through speaking engagements at industry events.
-Create high-quality educational content like videos, sample notebooks, datasets, tutorials, courseware, and blog posts.
- Expand and nurture the Databricks user community by organizing and growing meetups and user groups and providing support to data scientists, engineers, and analysts in online communities.
- Collaborate with product and engineering teams to share community learnings and influence product direction.
- Create and manage developer-focused programs to foster engagement and loyalty, creating a positive developer experience for data practitioners.
-Gather and analyze feedback from the community to drive continuous improvement of Databricks products and services.

What we look for:

- 5+ years experience as DevRel and as a software developer, solutions architect, or related profession
- Subject matter knowledge of solving data engineering problems at all scales.
- Active participation and recognized leadership in Databricks community forums, chats, and meetups
- Comprehensive understanding of the Databricks products and ecosystem
- Proven track record in nurturing developer communities, organizing user groups, and facilitating global meetups
- Exceptional communication skills, with a talent for articulating complex concepts through writing, teaching, videos, and public speaking
- Deep empathy for developer needs, with the ability to craft engaging experiences across tutorials, notebooks, and community platforms
- Adept at collaborating with cross-functional stakeholders to align community initiatives with product objectives
- Demonstrated ability to catalyze growth in both digital and in-person developer communities