Semester and Thesis Projects
The ETH AI Center offers a wide range of semester and thesis projects for students at ETH Zurich, as well as other universities. Please see the list below for projects that are currently available.
Are you a student? Check out our Semester and Thesis projects below!
Need help? If you have any questions, feel free to request them at
Language-Conditioned Interaction Trajectories Using Diffusion Algorithms for Robotic Manipulation
Train a diffusion model to predict 3D interaction trajectories from ego-centric images and task descriptions. The final model will be evaluated on a real robotic arm with five fingered hand.
Keywords
egocentric vision, diffusion models, representation learning, multimodal models, robotic manipulation
Labels
Master Thesis , ETH Zurich (ETHZ)
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2024-12-16
Organization Robotic Systems Lab
Hosts Zurbrügg René , Zurbrügg René
Topics Information, Computing and Communication Sciences
Safe RL for Robot Social Navigation
Developing a constrained RL framework for social navigation, emphasizing explicit safety constraints to reduce reliance on reward tuning.
Keywords
Navigation, Robot Planning, Reinforcement Learning, RL, Social Navigation
Labels
Master Thesis
Description
Goal
Contact Details
More information
Open this project... call_made
Published since: 2024-12-13 , Earliest start: 2025-01-01 , Latest end: 2025-12-31
Organization Spinal Cord Injury & Artificial Intelligence Lab
Hosts Alyassi Rashid , Alyassi Rashid , Alyassi Rashid
Topics Engineering and Technology
Bridging RL-based Robot Navigation & Crowd Simulation
Designing a crowd simulator for realistic human-robot interactions, enabling RL agent training in social navigation tasks.
Keywords
Navigation, Robot Planning, Reinforcement Learning, RL, Social Navigation.
Labels
Master Thesis
Description
Goal
Contact Details
More information
Open this project... call_made
Published since: 2024-12-12 , Earliest start: 2025-01-01 , Latest end: 2025-12-01
Organization Spinal Cord Injury & Artificial Intelligence Lab
Hosts Alyassi Rashid , Alyassi Rashid , Alyassi Rashid
Topics Engineering and Technology
Learn to Reach: Collision Aware End-Effector Path Planning & Tracking using Reinforcement Learning
Develop a method for collision aware reaching tasks using reinforcement learning and shape encodings of the environment
Keywords
Reinforcement Learning, Robotics, Perception, Machine Learning
Labels
Semester Project , Master Thesis , ETH Zurich (ETHZ)
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2024-12-09 , Earliest start: 2024-12-09 , Latest end: 2025-12-31
Applications limited to ETH Zurich
Organization Robotic Systems Lab
Hosts Zurbrügg René , Zurbrügg René , Portela Tifanny
Topics Information, Computing and Communication Sciences
Periodic Motion Priors for General Quadruped Locomotion Learning
In recent years, advancements in reinforcement learning have achieved remarkable success in quadruped locomotion tasks. Despite their similar structural designs, quadruped robots often require uniquely tailored reward functions for effective motion pattern development, limiting the transferability of learned behaviors across different models. This project proposes to bridge this gap by developing a unified, continuous latent representation of quadruped motions applicable across various robotic platforms. By mapping these motions onto a shared latent space, the project aims to create a versatile foundation that can be adapted to downstream tasks for specific robot configurations.
Keywords
representation learning, periodic autoencoders, policy modulating trajectory generators
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Miki Takahiro
Topics Information, Computing and Communication Sciences , Engineering and Technology
Lifelike Agility on ANYmal by Learning from Animals
The remarkable agility of animals, characterized by their rapid, fluid movements and precise interaction with their environment, serves as an inspiration for advancements in legged robotics. Recent progress in the field has underscored the potential of learning-based methods for robot control. These methods streamline the development process by optimizing control mechanisms directly from sensory inputs to actuator outputs, often employing deep reinforcement learning (RL) algorithms. By training in simulated environments, these algorithms can develop locomotion skills that are subsequently transferred to physical robots. Although this approach has led to significant achievements in achieving robust locomotion, mimicking the wide range of agile capabilities observed in animals remains a significant challenge. Traditionally, manually crafted controllers have succeeded in replicating complex behaviors, but their development is labor-intensive and demands a high level of expertise in each specific skill. Reinforcement learning offers a promising alternative by potentially reducing the manual labor involved in controller development. However, crafting learning objectives that lead to the desired behaviors in robots also requires considerable expertise, specific to each skill.
Keywords
learning from demonstrations, imitation learning, reinforcement learning
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Klemm Victor
Topics Information, Computing and Communication Sciences
Pushing the Limit of Quadruped Running Speed with Autonomous Curriculum Learning
The project aims to explore curriculum learning techniques to push the limits of quadruped running speed using reinforcement learning. By systematically designing and implementing curricula that guide the learning process, the project seeks to develop a quadruped controller capable of achieving the fastest possible forward locomotion. This involves not only optimizing the learning process but also ensuring the robustness and adaptability of the learned policies across various running conditions.
Keywords
curriculum learning, fast locomotion
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization Robotic Systems Lab
Hosts Li Chenhao , Bagatella Marco , Li Chenhao , Li Chenhao , Li Chenhao
Topics Engineering and Technology
Leveraging Human Motion Data from Videos for Humanoid Robot Motion Learning
The advancement in humanoid robotics has reached a stage where mimicking complex human motions with high accuracy is crucial for tasks ranging from entertainment to human-robot interaction in dynamic environments. Traditional approaches in motion learning, particularly for humanoid robots, rely heavily on motion capture (MoCap) data. However, acquiring large amounts of high-quality MoCap data is both expensive and logistically challenging. In contrast, video footage of human activities, such as sports events or dance performances, is widely available and offers an abundant source of motion data. Building on recent advancements in extracting and utilizing human motion from videos, such as the method proposed in WHAM (refer to the paper "Learning Physically Simulated Tennis Skills from Broadcast Videos"), this project aims to develop a system that extracts human motion from videos and applies it to teach a humanoid robot how to perform similar actions. The primary focus will be on extracting dynamic and expressive motions from videos, such as soccer player celebrations, and using these extracted motions as reference data for reinforcement learning (RL) and imitation learning on a humanoid robot.
Labels
Master Thesis
Description
Goal
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Kaufmann Manuel , Li Chenhao , Li Chenhao , Kaufmann Manuel , Li Chenhao
Topics Engineering and Technology
Learning Real-time Human Motion Tracking on a Humanoid Robot
Humanoid robots, designed to mimic the structure and behavior of humans, have seen significant advancements in kinematics, dynamics, and control systems. Teleoperation of humanoid robots involves complex control strategies to manage bipedal locomotion, balance, and interaction with environments. Research in this area has focused on developing robots that can perform tasks in environments designed for humans, from simple object manipulation to navigating complex terrains. Reinforcement learning has emerged as a powerful method for enabling robots to learn from interactions with their environment, improving their performance over time without explicit programming for every possible scenario. In the context of humanoid robotics and teleoperation, RL can be used to optimize control policies, adapt to new tasks, and improve the efficiency and safety of human-robot interactions. Key challenges include the high dimensionality of the action space, the need for safe exploration, and the transfer of learned skills across different tasks and environments. Integrating human motion tracking with reinforcement learning on humanoid robots represents a cutting-edge area of research. This approach involves using human motion data as input to train RL models, enabling the robot to learn more natural and human-like movements. The goal is to develop systems that can not only replicate human actions in real-time but also adapt and improve their responses over time through learning. Challenges in this area include ensuring real-time performance, dealing with the variability of human motion, and maintaining stability and safety of the humanoid robot.
Keywords
real-time, humanoid, reinforcement learning, representation learning
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts He Junzhe , Li Chenhao , Li Chenhao
Topics Information, Computing and Communication Sciences
Continuous Skill Learning with Fourier Latent Dynamics
In recent years, advancements in reinforcement learning have achieved remarkable success in teaching robots discrete motor skills. However, this process often involves intricate reward structuring and extensive hyperparameter adjustments for each new skill, making it a time-consuming and complex endeavor. This project proposes the development of a skill generator operating within a continuous latent space. This innovative approach contrasts with the discrete skill learning methods currently prevalent in the field. By leveraging a continuous latent space, the skill generator aims to produce a diverse range of skills without the need for individualized reward designs and hyperparameter configurations for each skill. This method not only simplifies the skill generation process but also promises to enhance the adaptability and efficiency of skill learning in robotics.
Keywords
representation learning, periodic autoencoders, learning from demonstrations, policy modulating trajectory generators
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Rudin Nikita
Topics Information, Computing and Communication Sciences , Engineering and Technology
Universal Humanoid Motion Representations for Expressive Learning-based Control
Recent advances in physically simulated humanoids have broadened their application spectrum, including animation, gaming, augmented and virtual reality (AR/VR), and robotics, showcasing significant enhancements in both performance and practicality. With the advent of motion capture (MoCap) technology and reinforcement learning (RL) techniques, these simulated humanoids are capable of replicating extensive human motion datasets, executing complex animations, and following intricate motion patterns using minimal sensor input. Nevertheless, generating such detailed and naturalistic motions requires meticulous motion data curation and the development of new physics-based policies from the ground up—a process that is not only labor-intensive but also fraught with challenges related to reward system design, dataset curation, and the learning algorithm, which can result in unnatural motions. To circumvent these challenges, researchers have explored the use of latent spaces or skill embeddings derived from pre-trained motion controllers, facilitating their application in hierarchical RL frameworks. This method involves training a low-level policy to generate a representation space from tasks like motion imitation or adversarial learning, which a high-level policy can then navigate to produce latent codes that represent specific motor actions. This approach promotes the reuse of learned motor skills and efficient action space sampling. However, the effectiveness of this strategy is often limited by the scope of the latent space, which is traditionally based on specialized and relatively narrow motion datasets, thus limiting the range of achievable behaviors. An alternative strategy involves employing a low-level controller as a motion imitator, using full-body kinematic motions as high-level control signals. This method is particularly prevalent in motion tracking applications, where supervised learning techniques are applied to paired input data, such as video and kinematic data. For generative tasks without paired data, RL becomes necessary, although kinematic motion presents challenges as a sampling space due to its high dimensionality and the absence of physical constraints. This necessitates the use of kinematic motion latent spaces for generative tasks and highlights the limitations of using purely kinematic signals for tasks requiring interaction with the environment or other agents, where understanding of interaction dynamics is crucial. We would like to extend the idea of creating a low-level controller as a motion imitator to full-body motions from real-time expressive kinematic targets.
Keywords
representation learning, periodic autoencoders
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Li Chenhao
Topics Information, Computing and Communication Sciences , Engineering and Technology
Humanoid Locomotion Learning and Finetuning from Human Feedback
In the burgeoning field of deep reinforcement learning (RL), agents autonomously develop complex behaviors through a process of trial and error. Yet, the application of RL across various domains faces notable hurdles, particularly in devising appropriate reward functions. Traditional approaches often resort to sparse rewards for simplicity, though these prove inadequate for training efficient agents. Consequently, real-world applications may necessitate elaborate setups, such as employing accelerometers for door interaction detection, thermal imaging for action recognition, or motion capture systems for precise object tracking. Despite these advanced solutions, crafting an ideal reward function remains challenging due to the propensity of RL algorithms to exploit the reward system in unforeseen ways. Agents might fulfill objectives in unexpected manners, highlighting the complexity of encoding desired behaviors, like adherence to social norms, into a reward function. An alternative strategy, imitation learning, circumvents the intricacies of reward engineering by having the agent learn through the emulation of expert behavior. However, acquiring a sufficient number of high-quality demonstrations for this purpose is often impractically costly. Humans, in contrast, learn with remarkable autonomy, benefiting from intermittent guidance from educators who provide tailored feedback based on the learner's progress. This interactive learning model holds promise for artificial agents, offering a customized learning trajectory that mitigates reward exploitation without extensive reward function engineering. The challenge lies in ensuring the feedback process is both manageable for humans and rich enough to be effective. Despite its potential, the implementation of human-in-the-loop (HiL) RL remains limited in practice. Our research endeavors to significantly lessen the human labor involved in HiL learning, leveraging both unsupervised pre-training and preference-based learning to enhance agent development with minimal human intervention.
Keywords
reinforcement learning from human feedback, preference learning
Labels
Master Thesis
Description
Goal
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Chen Xin , Li Chenhao
Topics Information, Computing and Communication Sciences , Engineering and Technology
Online Safe Locomotion Learning in the Wild
Reinforcement learning (RL) can potentially solve complex problems in a purely data-driven manner. Still, the state-of-the-art in applying RL in robotics, relies heavily on high-fidelity simulators. While learning in simulation allows to circumvent sample complexity challenges that are common in model-free RL, even slight distribution shift ("sim-to-real gap") between simulation and the real system can cause these algorithms to easily fail. Recent advances in model-based reinforcement learning have led to superior sample efficiency, enabling online learning without a simulator. Nonetheless, learning online cannot cause any damage and should adhere to safety requirements (for obvious reasons). The proposed project aims to demonstrate how existing safe model-based RL methods can be used to solve the foregoing challenges.
Keywords
safe mode-base RL, online learning, legged robotics
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao
Topics Engineering and Technology
Autonomous Curriculum Learning for Increasingly Challenging Tasks
While the history of machine learning so far largely encompasses a series of problems posed by researchers and algorithms that learn their solutions, an important question is whether the problems themselves can be generated by the algorithm at the same time as they are being solved. Such a process would in effect build its own diverse and expanding curricula, and the solutions to problems at various stages would become stepping stones towards solving even more challenging problems later in the process. Consider the realm of legged locomotion: Training a robot via reinforcement learning to track a velocity command illustrates this concept. Initially, tracking a low velocity is simpler due to algorithm initialization and environmental setup. By manually crafting a curriculum, we can start with low-velocity targets and incrementally increase them as the robot demonstrates competence. This method works well when the difficulty correlates clearly with the target, as with higher velocities or more challenging terrains. However, challenges arise when the relationship between task difficulty and control parameters is unclear. For instance, if a parameter dictates various human dance styles for the robot to mimic, it's not obvious whether jazz is easier than hip-hop. In such scenarios, the difficulty distribution does not align with the control parameter. How, then, can we devise an effective curriculum? In the conventional RSL training setting for locomotion over challenging terrains, there is also a handcrafted learning schedule dictating increasingly hard terrain levels but unified with multiple different types. With a smart autonomous curriculum learning algorithm, are we able to overcome separate terrain types asynchronously and thus achieve overall better performance or higher data efficiency?
Keywords
curriculum learning, open-ended learning, self-evolution, progressive task solving
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization Robotic Systems Lab
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Bagatella Marco , Li Chenhao
Topics Engineering and Technology
Humanoid Locomotion Learning with Human Motion Priors
Humanoid robots, designed to replicate human structure and behavior, have made significant strides in kinematics, dynamics, and control systems. Research aims to develop robots capable of performing tasks in human-centric settings, from simple object manipulation to navigating complex terrains. Reinforcement learning (RL) has proven to be a powerful method for enabling robots to learn from their environment, enhancing their performance over time without explicit programming for every possible scenario. In the realm of humanoid robotics, RL is used to optimize control policies, adapt to new tasks, and improve the efficiency and safety of human-robot interactions. However, one of the primary challenges is the high dimensionality of the action space, where handcrafted reward functions fall short of generating natural, lifelike motions. Incorporating motion priors into the learning process of humanoid robots addresses these challenges effectively. Motion priors can significantly reduce the exploration space in RL, leading to faster convergence and reduced training time. They ensure that learned policies prioritize stability and safety, reducing the risk of unpredictable or hazardous actions. Additionally, motion priors guide the learning process towards more natural, human-like movements, improving the robot's ability to perform tasks intuitively and seamlessly in human environments. Therefore, motion priors are crucial for efficient, stable, and realistic humanoid locomotion learning, enabling robots to better navigate and interact with the world around them.
Keywords
motion priors, humanoid, reinforcement learning, representation learning
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao
Topics Information, Computing and Communication Sciences
Learning World Models for Legged Locomotion
Model-based reinforcement learning learns a world model from which an optimal control policy can be extracted. Understanding and predicting the forward dynamics of legged systems is crucial for effective control and planning. Forward dynamics involve predicting the next state of the robot given its current state and the applied actions. While traditional physics-based models can provide a baseline understanding, they often struggle with the complexities and non-linearities inherent in real-world scenarios, particularly due to the varying contact patterns of the robot's feet with the ground. The project aims to develop and evaluate neural network-based models for predicting the dynamics of legged environments, focusing on accounting for varying contact patterns and non-linearities. This involves collecting and preprocessing data from various simulation environment experiments, designing neural network architectures that incorporate necessary structures, and exploring hybrid models that combine physics-based predictions with neural network corrections. The models will be trained and evaluated on prediction autoregressive accuracy, with an emphasis on robustness and generalization capabilities across different noise perturbations. By the end of the project, the goal is to achieve an accurate, robust, and generalizable predictive model for the forward dynamics of legged systems.
Keywords
forward dynamics, non-smooth dynamics, neural networks, model-based reinforcement learning
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization Robotic Systems Lab
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Klemm Victor , Li Chenhao
Topics Engineering and Technology
Studying Generalization, Compositionality and Sample Efficiency on Reasoning Tasks
Humans excel at breaking down complex scenes and understanding how objects within them relate to each other. In recent years, AI systems designed to test reasoning have made significant progress, with top performers now matching human accuracy on certain tests [1,2]. However, there's still a considerable difference in how efficiently humans and AI learn new tasks. Humans can pick up new skills with remarkably few examples, which is due to the ability to compose and generalize previously acquired knowledge in novel contexts. The ARC Challenge [3] is a remarkable example of a visual reasoning challenge where sample efficiency, compositionality and generalization are of primordial importance, in such a way that SOTA AI systems still significantly lag behind average human performance (46% for SOTA AI vs 100% for humans). In this project, we take the ARC Challenge as an inspiration to study the generalization, compositional and sample efficiency capabilities of different AI model families, using pre-existing datasets. We are also particularly interested in the recently discussed Grokking [4] phenomenon, and eager to exemplify this on visual based tasks.
Keywords
Artificial Intelligence, Sequence Models, ARC Challenge, Grokking
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2024-08-28 , Earliest start: 2024-08-28 , Latest end: 2025-06-30
Organization ETH Competence Center - ETH AI Center
Hosts Amo Carmen , Taoudi Yassine
Topics Information, Computing and Communication Sciences