Skip to yearly menu bar Skip to main content




ICLR 2024 Career Website

Here we highlight career opportunities submitted by our Exhibitors, and other top industry, academic, and non-profit leaders. We would like to thank each of our exhibitors for supporting ICLR 2024. Opportunities can be sorted by job category, location, and filtered by any other field using the search box. For information on how to post an opportunity, please visit the help page, linked in the navigation bar above.

Search Opportunities

Redmond, Washington, United States

The Human-Centered and AI Experiences area at Microsoft Research in Redmond, WA is looking for a candidate to advance the state of the art in evaluating and understanding the responsible development of foundational models with a focus on real-world use of Artificial Intelligence.

Our area conducts influential research published at top-tier venues in AI and ML (including NeurIPS, ICML, AAAI, and FAccT) and works within Microsoft’s Responsible AI ecosystem to impact our AI-driven technologies such as Azure, Office, and Bing.

We seek a researcher with a demonstrated ability for technical work and a proven record of influential publications on Artificial Intelligence.

For this role you’d need keen interest in rigorous evaluation, understanding, and innovation on foundational models.

Research areas of particular interest for this team include, but are not limited to:

  • Reliability & robustness of AI systems, rigorous evaluation and benchmarking, advances in AI interpretability, bias and fairness, and safety in real-world deployments

Our group takes a holistic approach to studying foundational models that includes a variety of data modalities (language, vision, multi-modal, and structured data) and modern model architectures.

You: - should have proven expertise in many of these aspects or show that they are interested in generalizing their contributions to a variety of modalities and architectures.

  • are a team players who are interested in developing next-generation platforms and tools for Machine Learning as well as conducting state-of-the-art research.

  • must have finished or be about to finish their PhD.

We are interested in both recent graduates and those with post-PhD academic or industry experience. Priority will be given to candidates with a proven publication record in top-tier conferences, who have demonstrated the ability to develop original research agendas and perform hands-on research, and who work well in a collaborative and dynamic environment.

Microsoft’s mission is to empower every person and every organization on the planet to achieve more. As employees we come together with a growth mindset, innovate to empower others, and collaborate to realize our shared goals. Each day we build on our values of respect, integrity, and accountability to create a culture of inclusion where everyone can thrive at work and beyond.

In alignment with our Microsoft values, we are committed to cultivating an inclusive work environment for all employees to positively impact our culture every day.


Apply

Cambridge, Cambridgeshire, United Kingdom

For our Microsoft Research Cambridge, UK, location, we are seeking Principal Researcher (equivalent to Director level/(US) tenured associate professor) candidates in the area of Machine Learning for Gaming and AI. We encourage applications from all candidates who are excited to tackle challenges that arise in applications of machine learning approaches to modern video games, including but not limited to those with backgrounds in deep learning, foundation models, generative AI, machine learning for computer vision or multi-modal models, deep reinforcement learning, or scalable and robust machine learning.

This is an exceptional opportunity for an experienced machine learning researcher and leader in their respective field to expand an established research group to drive ambitious research in collaboration with a diverse, multi-disciplinary team. Key research challenges that we are currently tackling include, but are not limited to, evaluation of multi-modal generative AI models, generalization, predictive (world) models, imitation and reinforcement learning, and scaling to large-scale data and compute.

This posting will be active for 25 days. All candidates applying while the post is active will be considered on an equal basis. For more information about the post, questions can be sent by email to the hiring manager Katja Hofmann at katja.hofmann@microsoft.com.

Microsoft’s mission is to empower every person and every organization on the planet to achieve more, and we’re dedicated to this mission across every aspect of our company. Our culture is centered on embracing a growth mindset and encouraging teams and leaders to bring their best each day. Join us and help shape the future of the world.


Apply

Location: Canada (Toronto, Waterloo, Vancouver, Montreal)


What's the opportunity?

As a Senior Machine Learning Researcher, you’re looking to channel your love of playing with real-world data into industry-disrupting solutions. We're a lab that supports research on a wide variety of theoretical and applied machine learning projects. Working in our lab will grant you unique access to massive structured and unstructured datasets with the tools and resources necessary to build game-changing statistical models.

Being part of our team means you’ll also have the opportunity to publish original research in peer-reviewed academic journals and participate in conferences around the world, such as NeurIPS, ICLR, ICML, CVPR and more.

Your responsibilities include:

  • Developing novel AI solutions that facilitate impactful products;
  • Conducting original, publishable research by advancing the state-of-the-art in machine learning techniques;
  • Working with the development team to transfer research work into production;
  • Interpreting larger organizational needs and designing algorithmic solutions that can drive the next generation of banking experiences;
  • Identifying relevant new AI technologies as they become available, and disseminating them into the bank’s technology capabilities.

You’re our ideal candidate if you have:

  • A passion for solving open problems using data and algorithms;
  • A PhD in a sub-area of AI or demonstrated research track record by means of publications and/or sophisticated AI product experience;
  • Ability to formulate and drive a research project independently without close supervision;
  • Proficiency in Python and Deep Learning packages such as Tensorflow or PyTorch;
  • Demonstrated ability to deliver machine learning solutions following industry best practices in software engineering is an asset;
  • Experience with time series or event forecasting, especially with asynchronous data, is an asset.

Apply

Location: Canada (Toronto, Waterloo, Vancouver, Montreal)


What’s the opportunity?

Borealis AI is looking for an enthusiastic Senior Research Engineer who’s excited by the opportunity of being at the forefront of machine learning technology, and working on extremely challenging problems in the financial services industry. As a Senior Research Engineer, you’ll be part of a collaborative team delivering AI projects end to end – everything from scoping and defining the AI problem, to data pre-processing and exploration, to prototyping novel algorithmic solutions, to software implementations of machine learning-based products. At Borealis AI, you’ll be joining a team that works directly with leading researchers in machine learning, have access to rich and massive datasets, and offers the computational resources to support cutting-edge machine learning R&D.

Your responsibilities include:

  • Delivering cutting edge AI solutions as a technical lead throughout the research and product development lifecycle;
  • Partnering with Borealis AI’s product and engineering teams to ensure the seamless delivery of these products;
  • Collaborating with business stakeholders to define the AI problem and rapidly prototype a solution;
  • Engaging in applied research experimentation to build and improve upon algorithms and models used in Borealis AI's products;
  • Supporting projects with thorough documentation, design decisions, and technical advisory.

You're our ideal candidate if you have:

  • Experience leading cross-functional machine learning research initiatives in product-centric environments;
  • Involvement across the research and product development lifecycle, from prototyping to production, and engaging with multiple stakeholders to develop solutions that meet business needs;
  • Proficient in Python, data science tooling, and deep learning frameworks;
  • Publication record in relevant top tier AI/ML conferences a plus;
  • Experience with NLP, LLMs, or financial services domain a plus.

Apply

Our roles are hybrid, 2-3 days per week from either our Palo Alto or NYC offices. This link ( https://boards.greenhouse.io/lightningai ) takes you to all of our openings.


Apply

Locations London, Amsterdam, Belgrade

We are currently in search of Staff Machine Learning (ML) engineers with exceptional skill sets and a passion for large language models. Your invaluable contributions will play a pivotal role in shaping our advanced LLM stack.

In this position, your responsibility will be to: - Conduct experiments with innovative deep learning methods to improve our models. - Explore techniques to enhance of our pretraining and alignment data.

We expect you to have: - 5+ years experience in a Machine Learning or AI related role - Substantial industrial or academic experience with large language models and generative AI. - A profound understanding of the theoretical foundations of machine learning. - Familiarity with modern deep learning frameworks such as TensorFlow, PyTorch, or JAX. - Proficiency in contemporary software engineering approaches, including CI/CD, version control, and unit testing. - A commitment to maintaining extreme rigor in all job-related activities.

Additional Advantages: - Previous experience working with language models or other similar NLP technologies. - Track record of building and delivering products (not necessarily ML-related) in a dynamic startup-like environment. - Strong engineering skills, including experience in developing large distributed systems or high-load web services. - Open-source projects that showcase your engineering prowess.


Apply

At Capital One, we are creating trustworthy and reliable AI systems, changing banking for good. For years, Capital One has been leading the industry in using machine learning to create real-time, intelligent, automated customer experiences. From informing customers about unusual charges to answering their questions in real time, our applications of AI & ML are bringing humanity and simplicity to banking. We are committed to building world-class applied science and engineering teams and continue our industry leading capabilities with breakthrough product experiences and scalable, high-performance AI infrastructure. At Capital One, you will help bring the transformative power of emerging AI capabilities to reimagine how we serve our customers and businesses who have come to love the products and services we build.


Apply

At Capital One, we are creating trustworthy and reliable AI systems, changing banking for good. For years, Capital One has been leading the industry in using machine learning to create real-time, intelligent, automated customer experiences. From informing customers about unusual charges to answering their questions in real time, our applications of AI & ML are bringing humanity and simplicity to banking. We are committed to building world-class applied science and engineering teams and continue our industry leading capabilities with breakthrough product experiences and scalable, high-performance AI infrastructure. At Capital One, you will help bring the transformative power of emerging AI capabilities to reimagine how we serve our customers and businesses who have come to love the products and services we build.


Apply

Department of Biomedical Informatics, Harvard Medical School


Summary

We are seeking a highly motivated and skilled Postdoctoral Researcher to join the labs of Dr. Isaac Kohane and Dr. Chirag Patel at Harvard Medical School Department for Biomedical Informatics. This position offers an exciting opportunity to contribute to cutting-edge research in single-cell transcriptomics with a specific focus on aging. The successful candidate will work collaboratively with Dr. Kohane's lab, Dr. Chirag Patel's lab, and Dr. Lee Rubin's lab at Harvard's Stem Cell and Regenerative Biology Department. This collaborative effort aims to advance our understanding of the molecular mechanisms underlying aging and age-related diseases using state-of-the-art single cell transcriptomic technologies.

Job Responsibilities

  • Conduct experiments in the field of single cell transcriptomics to investigate the molecular changes associated with aging and age-related diseases.
  • Perform data analysis and interpretation of single cell transcriptomic datasets, including data preprocessing, quality control, differential expression analysis, and identification of cell subtypes and their gene expression profiles.
  • Collaborate closely with Dr. Isaac Kohane's lab, Dr. Chirag Patel's lab, and Dr. Lee Rubin's lab to integrate single cell transcriptomic data with other omics data types, such as genomics, epigenomics, and proteomics, to gain comprehensive insights into the aging process.
  • Contribute to the development and optimization of novel single cell transcriptomic methodologies and technologies for studying aging and age-related diseases.
  • Work on the application of computational and statistical approaches, including machine learning algorithms, for the analysis and interpretation of single cell transcriptomic data.
  • Publish research findings in high-impact scientific journals and present research outcomes at conferences and seminars.
  • Collaborate with other members of the research teams, including graduate students and technicians, providing guidance and support in their research projects related to single cell transcriptomics and aging.
  • Stay up-to-date with the latest advancements in single cell transcriptomics, aging biology, and computational approaches, actively participating in scientific discussions and journal clubs.

Requirements

  • Ph.D. in a relevant field, such as Computational Biology, Bioinformatics, or a related discipline with a track record of publishing high-quality research papers in the field.
  • Strong expertise and hands-on experience in single cell transcriptomics data including preprocessing pipelines and downstream analyses.
  • Research background in the biology of aging and age-related diseases.
  • Proficiency in R or Python.
  • Experience with computational and statistical approaches, including machine learning algorithms, for analyzing biological data.
  • Familiarity with integrating multi-omics datasets and applying computational approaches for integrative analysis is highly desirable.

Apply

To apply, please submit a cover letter, detailed curriculum vitae, a list of references, and copies of your most significant publications or preprints to p1@zaklab.org.


Apply

Department of Biomedical Informatics, Harvard Medical School


Summary

We are seeking a highly motivated and skilled Postdoctoral Fellow to join the lab of Dr. Isaac Kohane at Harvard Medical School Department for Biomedical Informatics. This position offers an exciting opportunity to contribute to cutting-edge research in medical text processing using Large Language Models (LLM). The successful candidate will play a pivotal role in advancing our understanding of medical text analysis and its applications in healthcare.

Job Responsibilities

  • Conduct research and development activities using Large Language Models (LLM) for medical text processing.
  • Collaborate with a multidisciplinary team of researchers to design and implement novel algorithms and models for medical text analysis.
  • Analyze large-scale medical datasets, extract relevant information, and develop innovative approaches to improve the accuracy and efficiency of medical text processing.
  • Evaluate and validate the performance of developed algorithms and models through rigorous experimentation and benchmarking.
  • Contribute to the publication of research findings in high-impact scientific journals and present research outcomes at conferences and seminars.
  • Assist in mentoring and supervising researches in our lab, providing guidance and support in their research projects related to medical text processing.

Requirements

  • Ph.D. in Bioinformatics, Computational Linguistics, or a related field with a track record of publishing high-quality research papers in the field.
  • Strong expertise and hands-on experience in using LLMs for medical text processing.
  • Background in medicine or a related healthcare field is highly desirable.
  • Proficient programming skills in Python and R or another relevant programming language commonly used in data science and natural language processing.
  • Experience with large-scale medical datasets and familiarity with biomedical ontologies and terminologies.

Apply

To apply, please submit a cover letter, detailed curriculum vitae, a list of references, and copies of your most significant publications or preprints to p2@zaklab.org.


Apply

Cupertino, CA

Etched is building the hardware for superintelligence. GPUs and TPUs are flexible AI chips that can run many kinds of models: CNNs, RNNs, LSTMs, and more. But today, almost all AI workloads, from ChatGPT to self-driving cars, are done on one model architecture: transformers. Using flexible AI chips for transformers is very inefficient: <5% of the transistors on an H100 are used for matrix multiplication!

Etched is building a single-purpose chip exclusively for transformer inference. We only support transformers, but in exchange our chips have an order of magnitude more throughput and lower latency than an H100. With Etched, you can build products that would be impossible with GPUs, like tree-of-thought agents and ultra-low-latency audio chat bots.

Etched is looking for exceptional ML applications engineers to join our team. Building model-specific silicon unlocks new capabilities (e.g. tree search and super low latency applications) - an ideal candidate for this role will help develop products and work with customers who are developing products that aren’t possible without our hardware.

This role will report to the VP of Software.

Responsibilities:

  • Provide input for engineers designing our integrations with current transformer-specific inference libraries, like TensorRT-LLM, TransformerEngine, Hugging Face TGI, and vLLM.
  • Help profile and understand where latency comes from in modern LLM serving stacks
  • Help customers create products that leverage the unique capabilities of model-specific silicon

Requirements:

  • Deeply creative and able to think from first principles
  • Good understanding of LLM architecture and how to use them to build applications
  • 1+ year(s) of work experience at a cloud provider, AI company, or LLM startup
  • Experience writing performant real-time code AND proficient in Python
  • Breadth of knowledge about current research on large language models

Desired qualifications:

  • Experience with semiconductor design and development
  • Experience with deep learning frameworks (such as PyTorch, Tensorflow)
  • Experience with deep learning runtimes (such as ONNX Runtime, TensorRT,...)
  • Experience with at least one of TensorRT, TensorRT-LLM, Transformer Engine, or vLLM
  • Experience training, tuning and deploying ML models for CV (ResNet,..), NLP (BERT, GPT), and/or Recommendation Systems (DLRM)

Benefits:

  • Competitive salary and equity package
  • Full medical, dental, and vision packages, with 100% of premium covered
  • Work with world-class people and state-of-the-art AIs everyday
  • Etched is committed to fair and equitable compensation practices. Compensation is determined based on your qualifications and experience. Compensation packages also include generous equity in Etched.

We are an equal opportunity employer and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability or other legally protected statuses.


Apply

Cupertino, CA

Etched is building the hardware for superintelligence. GPUs and TPUs are flexible AI chips that can run many kinds of models: CNNs, RNNs, LSTMs, and more. But today, almost all AI workloads, from ChatGPT to self-driving cars, are done on one model architecture: transformers. Using flexible AI chips for transformers is very inefficient: <5% of the transistors on an H100 are used for matrix multiplication!

Etched is building a single-purpose chip exclusively for transformer inference. We only support transformers, but in exchange our chips have an order of magnitude more throughput and lower latency than an H100. With Etched, you can build products that would be impossible with GPUs, like tree-of-thought agents and ultra-low-latency audio chat bots.

We are seeking an experienced OpenCL/CUDA kernel developer with experience in embedded platforms to join our team. The ideal candidate will be responsible for developing and optimizing kernels for computing on specialized hardware platforms for parallel processing applications.

Responsibilities:

  • Design, developed, and optimized kernels for parallel processing for hardware acceleration.
  • Collaborate with domain experts to translate algorithms into efficient parallel processing.
  • Profile, analyze, and optimize kernel and overall performance to achieve maximum throughput and efficiency on acceleration hardware.
  • Implement and optimize data movement between CPU and various hardware blocks for effective utilization of parallel processing on an embedded platform. Implement synchronization mechanisms for data coordination between CPU and hardware acceleration, ensuring effective execution.
  • Develop and maintain test suites and benchmarks to validate kernel functionality, performance, and stability within embedded RTOS setup.

Requirements:

  • Experience in developing OpenCL or CUDA kernels for parallel computing.
  • 5+ years low level programming eg C/C++ in HPC or embedded environments.
  • Proficiency in programming languages such as C/C++ and expertise in OpenCL or CUDA.
  • Strong understanding of parallel programming, data structures, and optimization techniques suitable for embedded systems.
  • Experience with performance profiling and optimization tools for embedded systems.
  • Familiarity with memory management, data parallelism, and vectorization techniques.

Desired qualifications:

  • Experience working in hardware simulation/emulation environments.
  • Experience with implementing / serving LLMs

Benefits:

  • Competitive salary and equity package
  • Full medical, dental, and vision packages, with 100% of premium covered
  • Work with world-class people and state-of-the-art AIs everyday
  • Etched is committed to fair and equitable compensation practices. Compensation is determined based on your qualifications and experience. Compensation packages also include generous equity in Etched.

We are an equal opportunity employer and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability or other legally protected statuses.


Apply

Cupertino, CA

Etched is building the hardware for superintelligence. GPUs and TPUs are flexible AI chips that can run many kinds of models: CNNs, RNNs, LSTMs, and more. But today, almost all AI workloads, from ChatGPT to self-driving cars, are done on one model architecture: transformers. Using flexible AI chips for transformers is very inefficient: <5% of the transistors on an H100 are used for matrix multiplication!

Etched is building a single-purpose chip exclusively for transformer inference. We only support transformers, but in exchange our chips have an order of magnitude more throughput and lower latency than an H100. With Etched, you can build products that would be impossible with GPUs, like tree-of-thought agents and ultra-low-latency audio chat bots.

Etched is looking for an ML Research Scientist to help our customers co-design models and highly specialized microchips. We believe that as the costs of the largest LLMs continue to climb, model-hardware codesign will become essential for keeping inference affordable.

Responsibilities:

  • Design and implement deep learning architectures that will run efficiently on specialized silicon
  • Understand new advances in NLP and how they will work with our chip architecture
  • Accurately model performance of new transformer models on Etched’s specialized architecture
  • Provide feedback to the architecture and hardware teams about advances in kernel parallelism strategies, both within one system and across multiple

Requirements:

  • Experience designing architectures for large transformer models
  • PhD in Computer Science, Electrical and Computer Engineering, Mathematics, or a related scientific discipline, or equivalent experience
  • Deep and wide understanding of current research within large language models ​- Ability to program with Python or another scripting language
  • Passionate about AI scaling

Desired qualifications:

  • Proficiency with GPU programming.
  • Understanding of current techniques for efficient AI inference (including structured sparsity within memory, low precision floating point, and variants of attention).
  • Experience with industrial-scale training runs (that cost >$10M)

Benefits:

  • Competitive salary and equity package
  • Full medical, dental, and vision packages, with 100% of premium covered
  • Work with world-class people and state-of-the-art AIs everyday
  • Etched is committed to fair and equitable compensation practices. Compensation is determined based on your qualifications and experience. Compensation packages also include generous equity in Etched.

We are an equal opportunity employer and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability or other legally protected statuses.


Apply

Location: The School of Informatics at the University of Edinburgh, Edinburgh, UK

Deadline: 14th May 2024


We are looking for a postdoc to undertake novel research at the efficient deep neural network training in the School of Informatics at the University of Edinburgh.

The Topic The goal of this project is to develop novel techniques for training deep neural networks efficiently. This will involve generating data for training, sampling informative samples and designing data-efficient architectures for various vision models, and also vision-language models.

Project Team The candidate will be based in Visual Computing Group (https://groups.inf.ed.ac.uk/vico/) and will be supervised by Dr. Hakan Bilen. There will be many collaborative opportunities within the Vision group (https://groups.inf.ed.ac.uk/vision) as well as across the School of Informatics (which includes faculty working on machine learning, computer vision, NLP, speech processing, and social computing) and as well as with the University of Oxford and Bristol.

Your skills and attributes for success - PhD degree (or nearing completion) - A strong background in machine learning and/or computer vision - Publications at top venues in CV or/and ML (e.g., CVPR, ECCV, ICCV, NeurIPS, ICLR, ICML) - Excellent programming skills in Python, PyTorch, Tensorflow - Ability to work independently and manage own academic research and associated activities. - Excellence in written and oral communication, analytical, and time management skills.

Location The School of Informatics was ranked #1 in the UK for research power in Computer Science and Informatics Times Higher Education ranking, based on the 2021 Research Excellence Framework (REF). We are one of the top six institutions in Europe for AI according to CSRankings, and the highest-ranked UK institution. The School is an active Unit in the European Laboratory for Learning and Intelligent Systems (ELLIS). Edinburgh is a highly attractive city to live in. It is walkable, enjoys proximity to nature, and features a world-class cultural scene. It was ranked 1st in the 2022 Time-Out Index of the best cities in the world.

How to apply Apply at https://elxw.fa.em3.oraclecloud.com/hcmUI/CandidateExperience/en/sites/CX_1001/job/10261 before 14th May 2024

This post is full-time (35 hours per week) and fixed term for 12 months (with possibility of extension by another 12 months).

Contact details for enquiries:
Dr. Hakan Bilen h.bilen@ed.ac.uk https://homepages.inf.ed.ac.uk/hbilen


Apply

Location San Francisco, California, United States


Cognizant AI Labs

Cognizant works with an incredible diversity of organizations across the globe, using AI to improve decision-making, robustness, forecasting, and growth at every level of operation.

Within Cognizant, Cognizant AI Labs serves as the center of excellence for pioneering AI research. The team works to develop novel approaches to solve both fundamental scientific problems and challenges from real-world applications.

The work done by Cognizant AI Labs serves to inspire and catalyze real-world applications implemented by Cognizant, and reciprocally, real-world challenges encountered in Cognizant’s diverse ecosystem of applications serve to inspire foundational research at Cognizant AI Labs.

Your role:

As a research scientist, you will work with the Cognizant AI Labs research team to develop novel approaches to solving fundamental scientific problems and challenges from real-world applications, using core technologies such as LLMs, evolutionary algorithms, and other machine learning and AI techniques. With an explicit focus on AI for Good applications alongside basic research, the team envisions a world where AI systems are safe, robust, sustainable, long-lived, and inspiring. The team's current research is focused around areas including, but not limited to:

  • AI for decision-making
  • AI orchestration
  • Trustworthy AI
  • Open-ended AI
  • Sequential/time-series domains
  • Multi-agent systems
  • ALife
  • LLMs
  • Evolutionary computation

Key Responsibilities:

  • Work with the Advanced AI Labs research team to develop original ideas that can contribute to the AI community
  • Design experiments and evaluation methodologies for testing these ideas
  • Implement novel algorithms and evaluation frameworks
  • Manage experiments, analyze results, and iterate rapidly
  • Communicate ideas and results to a larger audience
  • Publish papers based on this work
  • Advise AI engineers on the development of practical applications

Qualifications:

  • Should have a PhD in Computer Science or another technical field.
  • Passion for AI research and AI for Good.
  • 3+ years of experience in AI or ML research.
  • Publications at venues such as ICLR, NeurIPS, GECCO, ALife, AAAI, IJCAI, ICML, etc.
  • Strong implementation skills.
  • Experience with LLMs.
  • Strong problem-solving and analytic skills.
  • Strong attention to detail and ability to work independently.
  • Excellent verbal and written communication skills.

Important Note: The application URL may occasionally expire or show "position filled". No worries! We are still hiring! If you are interested in this position, or have any questions about this job posting, feel free to contact the hiring manager Xin Qiu (xin.qiu@cognizant.com) or Elliot Meyerson (elliot.meyerson@cognizant.com).


Apply