Skip to yearly menu bar Skip to main content


MLSys 2026 Career Opportunities

Here we highlight career opportunities submitted by our Exhibitors, and other top industry, academic, and non-profit leaders. We would like to thank each of our exhibitors for supporting MLSys 2026.

Search Opportunities

Experience Required: Entry-level (PhD Program) or Experienced (Postdoc, Faculty, Scientific Lab)

Education: PhD in Math, Science, Engineering and other relevant disciplines

The PDT team - a quantitative investment manager - is hiring new or recent PhD graduates and experienced researchers (postdoctoral fellows, faculty, scientific lab, finance industry) to create and improve proprietary trading models and strategies while working closely with a deep bench of senior researchers.

We have a successful record of hiring, challenging, and retaining talented researchers from diverse academic backgrounds. Individuals interested in conducting innovative research with real-world impact seem to enjoy and excel at the types of problems we like to solve.

PDT Researchers work in small, nimble teams where merit and contribution, not seniority, drive the discussion. We strive to foster an intellectually challenging environment that encourages collaboration and innovative ideas.

In our research-driven approach to the financial markets, our Chief Scientist oversees the group-wide research agenda, ensuring team members work on the most critical and interesting problems, with a focus on research rigor and standards.

This is a hybrid position and will require the person to work from our New York City office at minimum 3 days a week. Why join us?

PDT Partners has a 30+ year track record and a reputation for excellence. Our goal is to be the best quantitative investment manager in the world—measured by the quality of our products, not their size. PDT’s very high employee-retention rate speaks for itself. Our people are intellectually extraordinary and our community is close-knit, down-to-earth, and diverse.

Responsibilities: Work closely with senior researchers on a variety of trading strategies and research projects, with the opportunity to conduct independent research and originate research topics over time Contribute to the long-term success of our research-driven algorithmic trading business

Below is a list of skills and experiences we think are relevant. Even if you don’t think you’re a perfect match, we still encourage you to apply because we are committed to developing our people. Solid mathematical and analytical ability; exceptional problem-solving and modeling ability Research intuition Experience in programming (Python, R, Matlab, C++) Excellent communication and collaborative white board skills Meticulous and detail-oriented, and innately driven to understand issues deeply Experience with/interested in working with large data sets Self-motivated and highly-productive, with a strong sense of ownership and urgency Able to work collaboratively and productively with others Enjoy solving complex, difficult, real-world problems Entrepreneurial and creative
Finance knowledge is not required or expected

The salary range for this role is between $190,000 and $250,000. This range is not inclusive of any potential bonus amounts. Factors that may impact the agreed upon salary within the range for a particular candidate include years of experience, level of education obtained, skill set, and other external factors.

PRIVACY STATEMENT: For information on ways PDT may collect, use, and process your personal information, please see PDT’s privacy notices.

We’re looking for excellent developers to join our growing Software Engineering organization.  Our work is collaborative, and our hiring reflects that. Interviewing at PDT is focused less on filling a specific role, and more on finding great people who can build long-term, varied careers with us.

Software Engineers at PDT are responsible for building and maintaining the technology the enables all parts of the trading life cycle, including building the trading systems, risk controls and post-trade technologies.

We are looking for people that can add to a company that values creativity, energy and problems that are solved by collective thinking. We are focused, deliberate, but nimble. We want our people to have the freedom to assess and then solve the challenging problems they are faced with independence and agility. This gives them an opportunity to make a direct impact on our bottom line. For the right talent, PDT offers fantastic growth potential.  

This is a hybrid position and will require the person to work from our New York City office at minimum 3 days a week.

Why join us? 

PDT Partners has a stellar 30+ year track record and a reputation for excellence. Our goal is to be the best quantitative investment manager in the world—measured by the quality of our products, not their size.  PDT’s very high employee-retention rate speaks for itself.  Our people are intellectually extraordinary, and our community is close-knit, down-to-earth, and diverse.  

Responsibilities:

Partner with internal end-users to understand (and anticipate) new features and requirements, then engineer efficient and effective solutions.

Develop and maintain our proprietary software stack using C++, Python, and Java.

Identifying, assessing, deploying the latest open-source and third-party software in both an on-prem and cloud environment.

Below is a list of skills and experiences we think are relevant. Even if you don’t think you’re a perfect match, we still encourage you to apply because we are committed to developing our people.

Significant experience programming in one or more of C++, Python, or Java. Experience working directly with users or clients, capturing requirements, and scoping Ability to participate in the design of complex software systems and select prudent and pragmatic technologies to fit the business objective.

Experience working with trading systems or financial data, working with low-latency systems, or working in a data science- or research-adjacent role a plus.

Education:

Bachelor’s or master’s degree in computer science

The salary range for this role is between $160,000 and $200,000. This range is not inclusive of any potential bonus amounts. Factors that may impact the agreed upon salary within the range for a particular candidate include years of experience, level of education obtained, skill set, and other external factors.

PRIVACY STATEMENT: For information on ways PDT may collect, use, and process your personal information, please see PDT’s privacy notices.

Location Santa Clara, California, USA or Toronto, Canada


Description At Lemurian Labs, we’re on a mission to bring the power of AI to everyone—without leaving a massive environmental footprint. We care deeply about the impact AI has on our society and planet, and we’re building a rock-solid foundation for its future, ensuring AI grows sustainably and responsibly. Because let’s face it, what good is innovation if it doesn’t help the world?

We are building a high-performance, portable compiler that lets developers “build once, deploy anywhere.” Yes, anywhere. We’re talking about seamless cross-platform compatibility, so you can train your models in the cloud, deploy them to the edge, and everything in between—all while optimizing for resource efficiency and scalability.

If the idea of sustainably scaling AI motivates you and you’re excited about making AI development both powerful and accessible, then we’d love to have you. Join us at Lemurian Labs, where you can have fun building the future—without leaving a mess behind.

The Role: We are looking for a Product Manager, Growth Lead who lives at the intersection of deep tech, community strategy, and product adoption. You will define how the market experiences the product by serving as the primary bridge between our product and the external developer and partner ecosystem, translating product value prop into developer adoption.

Here is what you will do: Drive Developer Adoption: Lead the "zero-to-one" experience for our AI stack. Architect the developer journey from discovery to deployment, ensuring our SDKs and libraries are not just powerful, but accessible. The Feedback Loop: Act as the "Customer Zero." Take insights from hackathons, partner integrations, neoclouds and open-source discord channels to help prioritize the product backlog. Technical Storytelling: Translate complexity to compelling technical narratives, whitepapers that resonate with ML engineers. Ecosystem Presence: Build high-trust relationships with developers and maintainers of open-source and Neo-cloud and Hyperscalar projects (e.g., vLLM, LangChain, Triton). Growth Engineering: Define and track key product metrics. Work with technical teams to build telemetry that informs growth strategy.

Essential Skills and Experience: Hybrid Background: 8+ years of experience blending Technical Product Management with Developer Relations, Solutions Engineering, or Growth roles in the AI/HPC space. Technical Fluency: You can read/write Python and understand lower-level concepts (GPU memory hierarchy, kernel fusion, latency vs. throughput) well enough to debate trade-offs with architects. Content Creation: A portfolio of technical writing (blogs, documentation) or public speaking that simplifies complex systems without dumbing them down. Execution: Proven track record of launching developer-facing products (APIs, SDKs, CLI tools) and measuring their success through adoption metrics, not just shipping dates. Education: BS/MS in Computer Science, Engineering, or equivalent practical experience.

Preferred Skills and Experience: Open Source Credibility: You have managed or contributed to high-growth repositories (>1k stars) or active Discord communities in the GenAI space. "Builder" DNA: You have personally built and deployed RAG pipelines or custom model serving endpoints in the last year. Network: You bring an existing network of relationships within the PyTorch, Hugging Face, Neoclouds or CNCF communities. Salary depends on experience and geographical location.

This salary range may be inclusive of several career levels and will be narrowed during the interview process based on a number of factors, such as the candidate’s experience, knowledge, skills, and abilities, as well as internal equity among our team.

Location Santa Clara, US or Toronto, Canada


Description At Lemurian Labs, we’re on a mission to bring the power of AI to everyone—without leaving a massive environmental footprint. We care deeply about the impact AI has on our society and planet, and we’re building a rock-solid foundation for its future, ensuring AI grows sustainably and responsibly. Because let’s face it, what good is innovation if it doesn’t help the world?

We are building a high-performance, portable compiler that lets developers “build once, deploy anywhere.” Yes, anywhere. We’re talking about seamless cross-platform compatibility, so you can train your models in the cloud, deploy them to the edge, and everything in between—all while optimizing for resource efficiency and scalability.

If the idea of sustainably scaling AI motivates you and you’re excited about making AI development both powerful and accessible, then we’d love to have you. Join us at Lemurian Labs, where you can have fun building the future—without leaving a mess behind.

Here is what you will do: - Design, develop, maintain and improve our multi-target runtime - Use the latest techniques in parallelization and partitioning to automate generation and exploit highly optimized kernels - Rapid prototyping and data driven exploration of new ideas - Benchmark and analyze the outputs produced by our optimizing compiler on target hardware - Work closely with our product team to understand the evolving needs of ML engineers and drive improvements in architecture - Build tools to collect and analyze performance bottleneck

Essential Skills and Experience: - BS degree in computer science, computer engineering, electrical engineering, or equivalent practical experience - 4+ years of experience working with compilers. - A deep understanding of asynchronous, concurrent programming. - 4+ years of experience with C/C++ (C++14 or newer). - An understanding of HW architecture (vector vs scalar registers and instructions, memory hierarchies). - Knowledge of operating system kernel development or hypervisor development.

Preferred Skills and Experience: - Masters or PhD degree in computer science, computer engineering, electrical engineering, or equivalent practical experience. - Experience developing or maintaining libraries like CUDA or ROCm. - Experience with GPU programming. - Experience with high performance computing (HPC). - Masters or PhD degree in computer science, or equivalent practical experience. - Knowledge of DL frameworks such as PyTorch, JAX or Triton. - Experience with programming large compute clusters.

Salary depends on experience and geographical location.

This salary range may be inclusive of several career levels and will be narrowed during the interview process based on a number of factors, such as the candidate’s experience, knowledge, skills, and abilities, as well as internal equity among our team.

Additional benefits for this role may include: equity, company bonus opportunities, medical, dental, and vision benefits; retirement savings plan; and supplemental wellness benefits.

Lemurian Labs ensures equal employment opportunity without discrimination or harassment based on race, color, religion, sex (including pregnancy, childbirth, or related medical conditions), sexual orientation, gender identity or expression, age, disability, national origin, marital or domestic/civil partnership status, genetic information, citizenship status, veteran status, or any other characteristic protected by law.

EOE

The Research Engineering team is dedicated to accelerating the velocity of machine learning research and expanding the exploration space for innovations at PDT. We partner with PDT’s quantitative researchers to design and build a state-of-the-art environment for testing ideas rapidly and efficiently.

Research at PDT requires significant compute, and as such, we are looking for a talented engineer with in-depth knowledge of ML techniques and DL ecosystem to help us build the infrastructure capable of supporting complex scientific research at scale.

This is a hybrid position and will require the person to work from our New York City office at minimum 3 days a week.

Why join us?  PDT Partners has a stellar 30+ year track record and a reputation for excellence. Our goal is to be the best quantitative investment manager in the world. PDT’s exceptional employee-retention rate speaks for itself. Our people are intellectually curious, collaborative, down-to-earth, and diverse.

Responsibilities:

Partner with the research team to understand future research directions and build the next generation of highly scalable infrastructure for alpha, signal, and portfolio construction.

Incorporate advancements in machine learning, hardware accelerators and high-performance computing to optimize research workflows.

Maintain, develop, and re-imagine the extensive internal research stack that continues to be a differentiating factor for PDT business. 

Optimize models for inference and use in real time trading systems.

Below is a list of skills and experiences we think are relevant. Even if you don’t think you’re a perfect match, we still encourage you to apply because we are committed to developing our people.

Experience with building infrastructure for training/fine-tuning large ML models.

Intellectual curiosity and a strong interest in solving difficult problems.

Exceptional programming skills and proficiency in identifying performance bottlenecks.

Experience with the python scientific stack and DL libraries (PyTorch, Tensorflow, etc.)

Experience with hardware accelerators.

Previous experience in Quant Finance is not required.

The salary range for this role is between $190,000 and $250,000. This range is not inclusive of any potential bonus amounts. Factors that may impact the agreed upon salary within the range for a particular candidate include years of experience, level of education obtained, skill set, and other external factors.

PRIVACY STATEMENT: For information on ways PDT may collect, use, and process your personal information, please see PDT’s privacy notices.

Location Santa Clara, California USA or Toronto, Canada


Description At Lemurian Labs, we’re on a mission to bring the power of AI to everyone—without leaving a massive environmental footprint. We care deeply about the impact AI has on our society and planet, and we’re building a rock-solid foundation for its future, ensuring AI grows sustainably and responsibly. Because let’s face it, what good is innovation if it doesn’t help the world?

We are building a high-performance, portable compiler that lets developers “build once, deploy anywhere.” Yes, anywhere. We’re talking about seamless cross-platform compatibility, so you can train your models in the cloud, deploy them to the edge, and everything in between—all while optimizing for resource efficiency and scalability.

If the idea of sustainably scaling AI motivates you and you’re excited about making AI development both powerful and accessible, then we’d love to have you. Join us at Lemurian Labs, where you can have fun building the future—without leaving a mess behind.

The Role We're looking for a Senior ML Performance Engineer to architect and lead our Performance Testing Platform from the ground up. You'll be the technical authority on how we measure, validate, and optimize the performance of large language models (Llama 3.2 70B, DeepSeek, and others) before and after compiler optimization on modern GPU architectures.

This is a high-impact role where you'll directly influence our product quality and our customers' success. You'll work at the intersection of ML systems, GPU architecture, and performance engineering—building the infrastructure that proves our compiler delivers real value.

Here is what you will do: Design and build a comprehensive performance testing platform for evaluating LLM inference workloads across GPU clusters Define and implement the benchmarking methodology, metrics, and test suites that measure latency, throughput, memory utilization, power consumption, and model accuracy Establish baseline performance for unoptimized models (Llama 3.2 70B, DeepSeek, etc.) and validate post-optimization improvements Develop automated testing pipelines for continuous performance validation across compiler releases and model updates Investigate performance bottlenecks using profiling tools (ROCm profilers, GPU traces, system-level monitoring) and work with the compiler team to drive optimizations Create dashboards and reporting that provide clear visibility into performance trends, regressions, and wins Collaborate cross-functionally with compiler engineers, ML engineers, and DevOps to ensure performance testing is integrated into our development workflow Document best practices for performance testing and optimization of ML workloads on GPU hardware

Essential Skills and Experience: BS degree in computer science, computer engineering, electrical engineering, or equivalent practical experience 7+ years of experience in performance engineering, benchmarking, or systems engineering roles Deep understanding of ML inference workloads, particularly transformer-based models and LLMs Hands-on experience with GPU programming and optimization (CUDA, ROCm, or similar) Strong programming skills in Python and C/C++ Proven track record of building performance testing infrastructure or benchmarking platforms from scratch Experience with ML frameworks (PyTorch, TensorFlow, ONNX Runtime, vLLM, TensorRT-LLM, etc.) Proficiency with profiling and debugging tools for GPU workloads Strong analytical skills with the ability to design experiments, analyze results, and communicate findings clearly Experience with CI/CD systems and test automation frameworks

We’re looking for an exceptional Performance Engineer to join our growing technology organization. Interviewing at PDT is intentionally focused on finding great people who can build long-term, impactful careers with us.

Performance Engineers at PDT are responsible for deeply understanding and optimizing the systems that enable our trading strategies at scale. You will work at the intersection of software, systems, and hardware to analyze performance, drive infrastructure efficiency, and free up critical compute capacity. Your work directly amplifies researcher velocity and scales our core models, creating massive impact through both cost savings and accelerated innovation. You'll thrive at PDT if you love open-ended problems, diving into GPU optimization and system optimization/design, and are excited to take your discoveries all the way to production at scale.

This is a hybrid position and will require the person to work from our New York City office at a minimum of 3 days a week.

Why join us

PDT Partners has a stellar 30+ year track record and a reputation for excellence. Our goal is to be the best quantitative investment manager in the world, measured by the quality of our products, not their size. PDT’s very high employee-retention rate speaks for itself. Our people are intellectually extraordinary, and our community is close-knit, down-to-earth, and diverse.

Key Responsibilities

Analyze and understand system performance to enhance researcher throughput and velocity.

Focus on infrastructure/system-level efficiency, working across Python, PyTorch, OS, networking, storage, and CPU/GPU layers to optimize compute resource utilization

Read and understand software layers, providing suggestions/PRs that optimize parts of codebases.

Free up capacity and reduce costs by improving computational efficiency

Support scaling of core models by ensuring efficient implementation

Propose and implement systems to improve performance telemetry

Conduct proof-of-concept (PoC) evaluations and contribute to system design

Identify and act on optimization opportunities across the stack

Below is a list of skills and experiences we think are relevant. Even if you don’t think you’re a perfect match, we still encourage you to apply because we are committed to developing our people.

Strong proficiency in Linux and its associated performance engineering toolset.

Experience with PyTorch, GPUs and CUDA for optimization.

Deep understanding and appreciation of what happens at the hardware-software interface.

Versatile engineering mindset: ability to learn quickly, tackle diverse challenges, and adapt.

Skills in coding, micro-optimization, and understanding multiple programming languages.

Ability to analyze performance without being solely focused on heads-down optimization.

The salary range for this role is between $195,000 and $225,000. This range is not inclusive of any potential bonus amounts. Factors that may impact the agreed upon salary within the range for a particular candidate include years of experience, level of education obtained, skill set, and other external factors.

PRIVACY STATEMENT: For information on ways PDT may collect, use, and process your personal information, please see PDT’s privacy notices.

Location Santa Clara, California US or Toronto, Canada


Description At Lemurian Labs, we’re on a mission to bring the power of AI to everyone—without leaving a massive environmental footprint. We care deeply about the impact AI has on our society and planet, and we’re building a rock-solid foundation for its future, ensuring AI grows sustainably and responsibly. Because let’s face it, what good is innovation if it doesn’t help the world?

We are building a high-performance, portable compiler that lets developers “build once, deploy anywhere.” Yes, anywhere. We’re talking about seamless cross-platform compatibility, so you can train your models in the cloud, deploy them to the edge, and everything in between—all while optimizing for resource efficiency and scalability.

If the idea of sustainably scaling AI motivates you and you’re excited about making AI development both powerful and accessible, then we’d love to have you. Join us at Lemurian Labs, where you can have fun building the future—without leaving a mess behind.

Here is what you will do: - Design, develop, maintain and improve our heterogeneous AI compiler. - Design and implement new capabilities in our compiler based on our novel compiler architecture. - Propose improvements to and expansions of our novel compiler architecture with respect to new advancements in machine learning model architectures and hardware. - Use the latest techniques in parallelization and partitioning to automate generation and exploit highly optimized kernels. - Generate and use performance data to identify opportunities and drive improvements. - Work with our product team to understand the evolving needs of ML engineers and drive improvements in architecture.

Essential Skills and Experience: - BS degree in computer science, computer engineering, electrical engineering, or equivalent practical experience - 4+ years of experience working with compilers. - Very strong knowledge of compiler algorithms and data structures. - Experience and interest in low level code generation, object file manipulation and target specific optimizations - 4+ years of experience with C/C++ - Strong written and oral communication, and able to write clear and concise documentation - Team first attitude - Detail oriented

Preferred Skills and Experience: - Masters or PhD degree in computer science, computer engineering, electrical engineering, or equivalent practical experience. - Knowledge of traditional compiler techniques; instruction selection, register allocation and traditional analysis like dominance, def-use et al. - Knowledge of calling conventions and APIs, linking and relocations. - Working knowledge of LLVM. - Experience with loop optimizations (vectorization, unrolling, fusion, parallelization, etc). - Experience with machine learning workloads and their demands on hardware.

Salary depends on experience and geographical location.

This salary range may be inclusive of several career levels and will be narrowed during the interview process based on a number of factors, such as the candidate’s experience, knowledge, skills, and abilities, as well as internal equity among our team.

Additional benefits for this role may include: equity, company bonus opportunities, medical, dental, and vision benefits; retirement savings plan; and supplemental wellness benefits.

Lemurian Labs ensures equal employment opportunity without discrimination or harassment based on race, color, religion, sex (including pregnancy, childbirth, or related medical conditions), sexual orientation, gender identity or expression, age, disability, national origin, marital or domestic/civil partnership status, genetic information, citizenship status, veteran status, or any other characteristic protected by law.

EOE

Location Santa Clara, California, USA or Toronto, Canada


Description At Lemurian Labs, we’re on a mission to bring the power of AI to everyone—without leaving a massive environmental footprint. We care deeply about the impact AI has on our society and planet, and we’re building a rock-solid foundation for its future, ensuring AI grows sustainably and responsibly. Because let’s face it, what good is innovation if it doesn’t help the world?

We are building a high-performance, portable compiler that lets developers “build once, deploy anywhere.” Yes, anywhere. We’re talking about seamless cross-platform compatibility, so you can train your models in the cloud, deploy them to the edge, and everything in between—all while optimizing for resource efficiency and scalability.

If the idea of sustainably scaling AI motivates you and you’re excited about making AI development both powerful and accessible, then we’d love to have you. Join us at Lemurian Labs, where you can have fun building the future—without leaving a mess behind.

About the Role As the founding member of our Developer Experience (DevX) team, you will be instrumental in shaping how engineers interact with our compiler infrastructure. You'll build the tools that give developers deep visibility into system performance—from profiling and debugging capabilities to hardware introspection interfaces. Your work will bridge the gap between our core compiler technology and the engineers who use it, transforming complex system data into actionable insights.

This role sits at the intersection of systems programming and developer tooling. You'll work closely with our compiler engineers to surface server-side telemetry through intuitive client-side interfaces, ultimately creating a best-in-class development experience for our users.

Here is what you will do: Design and build developer tools for profiling, debugging, and performance introspection across our compiler stack. Create client-side tooling that transforms server-side compiler telemetry into clear, actionable information for engineers. Develop interfaces that expose hardware performance metrics, and interrupt data in meaningful ways. Build GPU debugging capabilities and visualization tools to help engineers understand execution on heterogeneous hardware. Define formats and protocols for debug information exchange, working with standard debugger formats (DWARF, JTAG) and object file formats (ELF, COFF). Collaborate with internal engineering teams to understand their needs and iterate on tooling, with a path toward external customer-facing tools.

Essential Skills and Experience: 3+ years of professional experience in systems-level software development. Strong proficiency in C++ with experience writing performance-critical code. Working knowledge of assembly language and low-level debugging techniques. Familiarity with debugger formats (DWARF, JTAG) and object file formats (ELF, COFF). Understanding of profiling methodologies and performance analysis tools. Ability to work on-site at our Toronto or Santa Clara office.

Preferred Skills and Experience: Experience with GPU programming and debugging (CUDA, ROCm, or similar). Experience with OS-level interfaces including I/O subsystems and interrupt handling. Background in compiler development or toolchain infrastructure. Experience building developer-facing tools or IDEs. Contributions to open-source debugging or profiling tools.

Salary depends on experience and geographical location.

This salary range may be inclusive of several career levels and will be narrowed during the interview process based on a number of factors, such as the candidate’s experience, knowledge, skills, and abilities, as well as internal equity among our team.

Additional benefits for this role may include: equity, company bonus opportunities, medical, dental, and vision benefits; retirement savings plan; and supplemental wellness benefits.

Lemurian Labs ensures equal employment op