NeurIPS 2025 Career Opportunities
Here we highlight career opportunities submitted by our Exhibitors, and other top industry, academic, and non-profit leaders. We would like to thank each of our exhibitors for supporting NeurIPS 2025.
Search Opportunities
San Francisco
We are seeking a talented software engineer with generative AI experience who is deeply proficient with python and typescript to join our dynamic and growing team at Writer. As a key member of our engineering team, you will play a crucial role in building the genAI software. Your primary focus will be on developing a state-of-the-art platform that harnesses generative AI technologies and you will deliver seamless and scalable solutions. You will work closely with cross-functional teams to design, implement, and maintain features that enhance the user experience, drive product growth, establish best practices, and integrate cutting-edge AI capabilities.
Your responsibilities
- Design and develop robust and scalable generative AI services using Python and open source frameworks such as Writer Agent Builder, LangChain, and n8n.
- Implement responsive and user-friendly frontend interfaces, leveraging technologies like React, TypeScript, and modern web frameworks.
- Work with cloud platforms such as AWS, GCP, or Azure to deploy and scale applications.
- Develop and integrate high-performance, low-latency APIs for AI-driven features.
- Ensure code quality through testing, peer reviews, and continuous integration.
- Collaborate with the team to build, and maintain generative AI agents.
- Participate in architectural design discussions and promote engineering best practices.
- Continuously improve the application’s performance, scalability, and maintainability.
Is This You?
- 5+ years of experience in software engineering at expert level with Python
- Experience building web applications using FastAPI and Asyncio
- Experience building with generative AI applications in production environments.
- Expertise with microservices architecture and RESTful APIs.
- Solid understanding of database technologies such as PostgreSQL and vector databases as Elastic, Pinecone, Weaviate, or similar.
- Familiarity with cloud platforms (AWS, GCP, etc.) and containerized environments (Docker, Kubernetes).
- Familiarity with MCP, devtools, AI agents, or contributed to open source
- You are committed to writing clean, maintainable, and scalable code, following best practices in software development.
- You enjoy solving complex problems and continuously improving the performance and scalability of systems.
- You thrive in collaborative environments, working closely with cross-functional teams to build impactful features.
- Proven ability to help teams adopt technical best practices.
Position: Data Science Intern
Location: 660 5th Avenue, New York, NY
Viking Global Investors (“Viking”) is a global investment firm founded in 1999, managing over $53 billion in capital across public and private investments. With offices in Stamford, New York, Hong Kong, London, and San Francisco, Viking is registered with the U.S. Securities and Exchange Commission. For more information, visit www.vikingglobal.com.
Internship Opportunity
The Data Science Intern will collaborate with the Data Science team, Investment Analysts, and Data Engineers to analyze and expand Viking’s alternative data assets, generating actionable investment insights. This role is ideal for analytical, creative problem solvers eager to apply their data science skills to pressing research questions. Interns will work both independently and alongside quantitative professionals, with flexibility in duration, start dates, and full-time/part-time options.
Informational Webinar: October 30, 6:00–7:00pm ET
Register here
Responsibilities
- Develop and deliver predictive analytics on companies, sectors, and macroeconomic trends
- Generate investment insights from alternative data analysis
- Create methodologies to identify and evaluate private company investment opportunities
- Identify and assess new data sources
- Streamline data lifecycle, operating models, and processes
- Test and evaluate new technologies for the big data platform
- Build centralized, automated analyses and processes
- Share information and insights to support Viking’s research efforts
Qualifications
- Currently enrolled in a Master’s or PhD program (3rd year+) in Data Science, Economics, Finance, Statistics, or related quantitative fields
- Strong communication skills, with the ability to explain complex ideas to non-technical audiences
- Independent thinker, capable of leading research projects with partial supervision
- Proficient in Python, statistical libraries, SQL, BI tools (e.g., Tableau), and cloud technologies
- Sound judgment and big-picture perspective
- Passionate about research, proactive, and self-motivated
- Committed to excellence
Application
Submit your resume and a 1–2 page supplement describing a recent quantitative research project via the Viking career site.
Supplement must include:
- Research question
- Data used
- Approach and statistical methodologies
- Findings
- Computational environment (language, main libraries, etc.)
Application Deadline: November 11, 2025 (11:59 PM EST)
Interviews: Conducted virtually in December
Compensation & Benefits
- Base Salary Range (NYC): $175,000 – $250,000 annually
- Actual compensation determined by skill set, experience, education, and qualifications
Equal Opportunity Employer
Viking is an equal opportunity employer. For questions or accommodation requests, contact:
Viking Campus Recruiting Team
campusrecruiting@vikingglobal.com
New York
Software Developer
Technology is integral to virtually everything the D. E. Shaw group does, which is why we seek exceptional software developers with a range of quantitative and programming abilities. Members of our technical staff collaborate on challenging problems that directly impact the firm’s continued success, utilizing their excellent analytical, mathematical, and software design skills as well as some of the most advanced computing resources in the world. Software developers have the opportunity to be part of a collegial, collaborative, and engaging working environment.
What you'll do day-to-day
Specific responsibilities may include formulating statistical models for our computerized trading strategies, developing distributed systems to analyze and react to incoming data in real time, and creating tools for advanced mathematical modeling.
Who we're looking for
- Successful developers have traditionally been the top students in their programs and have extensive software development experience.
- We welcome outstanding candidates at all experience levels.
- The expected annual base salary for this position is 225,000USD. Our compensation and benefits package includes substantial variable compensation in the form of a year-end bonus, guaranteed in the first year of hire, a sign-on bonus, a relocation bonus, and benefits including medical and prescription drug coverage, 401(k) contribution matching, wellness reimbursement, family building benefits, and a charitable gift match program.
Amsterdam
Flow Traders is committed to leveraging the most recent advances in machine learning, computer science, and AI to generate value in the financial markets. We are looking for Quantitative Researchers to join this challenge.
As a Quantitative Researcher at Flow Traders, you are an expert in mathematics and statistics. You are passionate about translating challenging problems into equations and models, and have the ability to optimize them using cutting-edge computational techniques. You collaborate with a global team of researchers and engineers to design, build, and optimize our next generation of models and trading strategies.
Are you at the top of your quantitative, modeling, and coding game, and excited by the prospect of demonstrating these skills in competitive live markets? Then this opportunity is for you.
Palo Alto, CA
Position Description: Tesla’s AI team is pushing the frontier of real-world machine learning, building models that reason, predict, and act with human-level physical intelligence. We train and deploy large-scale ML systems powering products from Autopilot to Optimus.
As part of the Model Optimization group, you will work at the intersection of machine learning and systems, designing our most advanced models to run efficiently across Tesla’s diverse compute stack, from data centers to edge AI accelerators. You will design the model architecture and engineer algorithmic optimizations that make large-scale model inference fast, reliable, and hardware-aware.
Responsibilities: Design, train, and deploy large neural networks that run efficiently on heterogeneous hardware (GPU, CPU, Tesla’s in-house AI ASIC) Develop and integrate quantization, sparsity, pruning, and distillation techniques to improve inference performance Design inference algorithms that improve inference performance in terms of quantization and latency Profile and improve latency, throughput, and memory efficiency for large ML models across edge and cloud environments Collaborate with compiler and hardware engineers to co-design architectures for efficient real-time inference Design and implement custom GPU kernels (CUDA / OpenCL) to accelerate model operations and post-processing pipelines Conduct systematic benchmarking, scaling, and validation of inference performance across Tesla platforms
Requirements: Proven experience in scaling and optimizing inference for large ML models, particularly transformers or similar architectures Familiarity with quantization-aware training, model compression, and distillation for edge and real-time inference Proficiency with Python and C++ (modern standards 14/17/20) and deep learning frameworks such as PyTorch, TensorFlow, or JAX Strong understanding of computer systems and architecture, with experience deploying ML models on GPUs, TPUs, or NPUs Hands-on expertise with CUDA programming, low-level performance profiling, and compiler-level optimization (TensorRT, TVM, XLA) Experience collaborating with compiler/hardware engineers to bridge model and system-level optimization Excellent problem-solving skills and the ability to debug and tune high-performance inference workloads
UK
Research Engineer - Novel AI applications and Next Generation Hardware
Mission: You will join the hardware team with the goal of supporting novel application areas and AI modes beyond current use cases. Responsibilities include researching the evolving landscape of AI applications and models, analyzing underlying model architectures, and building implementations on Groq. Further responsibilities include analyzing mappings to existing and future hardware, modeling performance, and working cross-functionally with the hardware design team on novel hardware features e.g. functional units, numeric modes, interconnect, system integration, etc to unlock novel application areas for Groq. There will be opportunities to participate in a wider range of R&D activities, either internally or externally with key Groq partners.
Responsibilities & opportunities in this role: AI application and model research Performance modeling Cross-functional work with hardware and software teams Next generation hardware architecture development Support internal and outward-facing R&D
Ideal candidates have/are: Strong foundation in computer science Experience with AI models and applications Knowledge of LLMs and other Gen AI applications Strong foundation in computer architecture and computer arithmetic Python and common ML frameworks such as PyTorch & TensorFlow Experience with performance analysis / modelling Problem solving mindset
Nice to Have: Experience with scientific computing & HPC Experience in optimizing applications on specialized accelerators (GPU, FPGA, or other custom accelerators). Experience with compiler tools and MLIR. Experience in delivering complex projects in a fast-moving environment.
Attributes of a Groqster: Humility - Egos are checked at the door Collaborative & Team Savvy - We make up the smartest person in the room, together Growth & Giver Mindset - Learn it all versus know it all, we share knowledge generously Curious & Innovative - Take a creative approach to projects, problems, and design Passion, Grit, & Boldness - no limit thinking, fueling informed risk taking
Compensation: At Groq, a competitive base salary is part of our comprehensive compensation package, which includes equity and benefits. For this role, salary range is determined by your location, skills, qualifications, experience and internal benchmarks. Compensation for candidates outside the USA will be dependent on the local market.
This position may require access to technology and/or information subject to U.S. export control laws and regulations, as well as applicable local laws and regulations, including the Export Administration Regulations (EAR). To comply with these requirements, candidates for this role must meet all relevant export control eligibility criteria.
We are now looking for a Senior Research Scientist for Generative AI!
NVIDIA is searching for a world-class researcher in generative AI to join our research team. You will be conducting original research for generative AI applications, including image generation, video generation, 3D generation, and audio generation. You will be working with a team of world-class researchers eager to make great impacts with generative AI models. You will be building research prototypes and scaling them with large datasets and compute. After building prototypes that demonstrate the promise of your research, you will work with product teams to help them integrate your ideas into products.
What you'll be doing: Conduct original research in the space of generative AI
Implement and train large-scale generative AI models for various content creation applications
Collaborate with other research team members, a diverse set of internal product teams, and external researchers
Have a broader impact through the transfer of the technology you've developed to relevant product groups
What we need to see: Ph.D. in Computer Science/Engineering, Electrical Engineering, or a related field (or equivalent experience).
5+ years of relevant research experience.
Excellent collaboration and interpersonal skills
Excellent python/C++ programming skills
Great knowledge of common deep-learning frameworks
Experience in processing or curating large-scale datasets
Excellent knowledge of theory and practice of deep learning, computer vision, natural language processing, or computer graphics
Track record of research excellence or significant product development
NVIDIA is widely considered to be one of the technology world’s most desirable employers. We have some of the most forward-thinking and hardworking people in the world working for us. Are you a creative and autonomous research scientist with a genuine passion for advancing the state of AI? If so, we want to hear from you!
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 184,000 USD - 299,000 USD for Level 4, and 224,000 USD - 356,500 USD for Level 5.
You will also be eligible for equity and benefits.
NVIDIA is committed to fostering a diverse work environment and proud to be an equal opportunity employer. As we highly value diversity in our current and future employees, we do not discriminate (including in our hiring and promotion practices) on the basis of race, religion, color, national origin, gender, gender expression, sexual orientation, age, marital status, veteran status, disability status or any other characteristic protected by law.
Preference for on-site candidates in San Mateo, but remote possible.
BigHat is hiring a Principal ML Scientist. We've got an awesome high-throughput wetlab that pumps proprietary data into custom ETL and ML Ops infra to power our weekly design-build-train loop. Come solve hard-enough-to-be-fun problems in protein engineering in service of helping patients!
Noumenal Labs | Remote-friendly | Full-time
Noumenal’s Probabilistic Perception Lab builds systems capable of navigating outdoor environments through probabilistic spatial reasoning and structured uncertainty reduction. We are looking for a Research Engineer with deep experience in probabilistic inference, spatial AI, and structured generative models to drive applied breakthroughs in perception of outdoor environments. You will work closely with researchers, systems engineers, commercial software engineers, and roboticists to build models that integrate 3D geometry, scene composition, uncertainty, and adaptive inference grounded in generative representations. This role is ideal for someone who has operated at the intersection of probabilistic computing, 3D scene understanding, computational neuroscience, and machine learning research, with experience spanning both foundational research and scalable, applied engineering.
What You’ll Do
~ Develop and deploy probabilistic generative models for perception, scene understanding, and spatial reasoning (structured generative models, inverse graphics, Bayesian scene reconstruction) on hardware in a commercial product. ~ Build inference engines for SLAM, 3D reconstruction, object-centric scene modeling, and spatial world models, leveraging MCMC, variational inference, or novel structured inference techniques. ~ Design systems that combine topological, geometric, and probabilistic methods for robust representation of spatial and conceptual structure. ~ Lead and engage in directed engineering efforts to translate novel algorithms into performant systems suited for real-time or near–real-time perception. ~ Collaborate with researchers in probabilistic computing, robotics, and AI to prototype, test, and iterate on models using synthetic and real sensory data.
Required Skills
~ Experience building perception systems in robotics. ~ Ability to translate research concepts into robust, scalable engineering implementations. ~ Strong coding ability in Python and modern ML frameworks (PyTorch, JAX, or TensorFlow). ~ Expertise in probabilistic inference, structured generative models, or Bayesian approaches (MCMC, variational inference, factorized models, hierarchical generative models). ~ Experience in 3D perception and spatial AI, including at least one of: SLAM, object-centric modeling, structured scene representations, or probabilistic inverse graphics frameworks. ~ Commitment to open-source contributions and internal cross-lab collaborations.
Ideal Background
~ Experience with topological data analysis, geometric representations, or mathematical structure in learning systems (e.g., planning in latent spaces). ~ Strong mathematical background (geometry, topology, optimization, or probabilistic modeling). ~ Background working in interdisciplinary research groups (AI, neuroscience, robotics, mathematics). ~ Publications in machine learning, probabilistic modeling, computational neuroscience, or mathematical methods for perception.
What We Offer
~ Close collaboration with researchers in robotics, physics-inspired AI, and spatial intelligence. ~ Access to real-world data for 3D perception and inference experiments. ~ A remote-friendly environment, flexible work culture, competitive salary + equity.
The role We are seeking a highly skilled and customer-focused professional to join our team as a Cloud Solutions Architect specializing in Cloud infrastructure and MLOps. As a Cloud Solutions Architect, you will play a pivotal role in designing and implementing cutting-edge solutions for our clients, leveraging cloud technologies for ML/AI teams and becoming a trusted technical advisor for building their pipelines.
You’re welcome to work remotely from the US or Canada.
Your responsibilities will include: - Act as a trusted advisor to our clients, providing technical expertise and guidance throughout the engagement. Conduct PoC, workshops, presentations, and training sessions to educate clients on GPU cloud technologies and best practices. - Collaborate with clients to understand their business requirements and develop solution architecture that align with their needs: design and document Infrastructure as code solutions, documentation and technical how-tos in collaboration with support engineers and technical writers. - Help customers to optimize pipeline performance and scalability to ensure efficient utilization of cloud resources and services powered by Nebius AI. - Act as a single point of expertise of customer scenarios for product, technical support, marketing teams. - Assist to Marketing department efforts during events (Hackathons, conferences, workshops, webinars, etc.)
We expect you to have: - 5 - 10 + years of experience as a cloud solutions architect, system/network engineer, developer or a similar technical role with a focus on cloud computing - Strong hands-on experience with IaC and configuration management tools (preferably Terraform/Ansible), Kubernetes, skills of writing code in Python - Solid understanding of GPU computing practices for ML training and inference workloads, GPU software stack components, including drivers, libraries (e.g. CUDA, OpenCL) - Excellent communication skills - Customer-centric mindset
It will be an added bonus if you have: - Hands-on experience with HPC/ML orchestration frameworks (e.g. Slurm, Kubeflow) - Hands-on experience with deep learning frameworks (e.g. TensorFlow, PyTorch) - Solid understanding of cloud ML tools landscape from industry leaders (NVIDIA, AWS, Azure, Google)