Student Projects

Currently, the following student projects are available. Please contact the responsible supervisor and apply with your CV and transcripts.

In case you have project ideas related to any of these projects, take the opportunity and propose your own project!

We also offer Master Thesis and or exchange semesters at reknown universities around the world. Please contact us in case of interest!

Studies on Mechatronics

We offer students also to conduct their Studies on Mechatronics at our lab. In general, we recommend to do the Studies on Mechatronics in combination with the Bachelor Thesis, either as prepartory work the semester beofre or as extended study in parallel. If you want to do it independently, yiou can find prroposed projets also in the list below. Please directly apply with corresponsponding supervisor.

ETH Zurich uses SiROP to publish and search scientific projects. For more information visit sirop.org.

Reinforcement learning to reject large disturbances for a robotic arm

Quasi-direct drive robot arms experience high deflection in the presence of large end effector forces. This leads to stability issues when these forces change rapidly, such as when dropping a heavy object. Existing works use reinforcement learning to achieve end effector tracking[1] as well as force control[2]. In this project, you will extend upon this by developing a controller capable of tracking end effector poses in the presence of sudden, large changes in end effector disturbance. You will first deploy your controller on a standalone robotic arm and then work towards deployment on our ANYmal-based bimanual mobile manipulator. [1] Martín-Martín et al., Variable Impedance Control in End-Effector Space: An Action Space for Reinforcement Learning in Contact-Rich Tasks (IROS 2019) [2] Portela et al., Learning Force Control for Legged Manipulation (ICRA 2024)

Keywords

reinforcement learning, robot arm, mobile manipulation, force disturbance

Labels

Semester Project , Bachelor Thesis

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-03-07 , Earliest start: 2025-03-10

Organization Robotic Systems Lab

Hosts Mittal Mayank , Fischer Oliver

Topics Engineering and Technology

Novel Winch Control for Robotic Climbing

While legged robots have demonstrated impressive locomotion performance in structured environments, challenges persist in navigating steep natural terrain and loose, granular soil. These challenges extend to extraterrestrial environments and are relevant to future lunar, martian, and asteroidal missions. In order to explore the most extreme terrains, a novel winch system has been developed for the ANYmal robot platform. The winch could potentially be used as a fail-safe device to prevent falls during unassisted traverses of steep terrain, as well as an added driven degree of freedom for assisted ascending and descending of terrain too steep for unassisted traversal. The goal of this project is to develop control policies that utilize this new hardware and enable further climbing robot research.

Keywords

Robot, Space, Climbing, Winch, Control

Labels

Semester Project , Master Thesis , ETH Zurich (ETHZ)

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-03-05 , Earliest start: 2024-10-07

Organization Robotic Systems Lab

Hosts Vogel Dylan

Topics Information, Computing and Communication Sciences , Engineering and Technology

Beyond Value Functions: Stable Robot Learning with Monte-Carlo GRPO

Robotics is dominated by on-policy reinforcement learning: the paradigm of training a robot controller by iteratively interacting with the environment and maximizing some objective. A crucial idea to make this work is the Advantage Function. On each policy update, algorithms typically sum up the gradient log probabilities of all actions taken in the robot simulation. The advantage function increases or decreases the probabilities of these taken actions by comparing their “goodness” versus a baseline. Current advantage estimation methods use a value function to aggregate robot experience and hence decrease variance. This improves sample efficiency at the cost of introducing some bias. Stably training large language models via reinforcement learning is well-known to be a challenging task. A line of recent work [1, 2] has used Group-Relative Policy Optimization (GRPO) to achieve this feat. In GRPO, a series of answers are generated for each query-answer pair. The advantage is calculated based on a given answer being better than the average answer to the query. In this formulation, no value function is required. Can we adapt GRPO towards robot learning? Value Functions are known to cause issues in training stability [3] and a result in biased advantage estimates [4]. We are in the age of GPU-accelerated RL [5], training policies by simulating thousands of robot instances simultaneously. This makes a new monte-carlo (MC) approach towards RL timely, feasible and appealing. In this project, the student will be tasked to investigate the limitations of value-function based advantage estimation. Using GRPO as a starting point, the student will then develop MC-based algorithms that use the GPU’s parallel simulation capabilities for stable RL training for unbiased variance reduction while maintaining a competitive wall-clock time.

Keywords

Robot Learning, Reinforcement Learning, Monte Carlo RL, GRPO, Advantage Estimation

Labels

Semester Project , Bachelor Thesis , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-03-05

Organization Robotic Systems Lab

Hosts Klemm Victor

Topics Information, Computing and Communication Sciences , Engineering and Technology , Behavioural and Cognitive Sciences

Manipulation beyond Single End-Effector

The goal of the project is to extend our prior works to make ANYmal with an arm use its different end-effectors for whole-body mobile manipulation.

Keywords

reinforcement learning, robotics, perception, robot learning

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-03-03 , Earliest start: 2025-03-17

Organization Robotic Systems Lab

Hosts Mittal Mayank

Topics Information, Computing and Communication Sciences , Behavioural and Cognitive Sciences

Volumetric Bucket-Fill Estimation

Gravis Robotics is an ETH spinoff from the Robotic Systems Lab (RSL) working on the automation of heavy machinery (https://gravisrobotics.com/). In this project, you will be working with the Gravis team to develop a perceptive bucket-fill estimation system. You will conduct your project at Gravis under joint supervision from RSL.

Keywords

Autonomous Excavation

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-02-28 , Earliest start: 2025-01-01 , Latest end: 2026-01-01

Organization Robotic Systems Lab

Hosts Egli Pascal Arturo

Topics Engineering and Technology

Leveraging Human Motion Data from Videos for Humanoid Robot Motion Learning

The advancement in humanoid robotics has reached a stage where mimicking complex human motions with high accuracy is crucial for tasks ranging from entertainment to human-robot interaction in dynamic environments. Traditional approaches in motion learning, particularly for humanoid robots, rely heavily on motion capture (MoCap) data. However, acquiring large amounts of high-quality MoCap data is both expensive and logistically challenging. In contrast, video footage of human activities, such as sports events or dance performances, is widely available and offers an abundant source of motion data. Building on recent advancements in extracting and utilizing human motion from videos, such as the method proposed in WHAM (refer to the paper "Learning Physically Simulated Tennis Skills from Broadcast Videos"), this project aims to develop a system that extracts human motion from videos and applies it to teach a humanoid robot how to perform similar actions. The primary focus will be on extracting dynamic and expressive motions from videos, such as soccer player celebrations, and using these extracted motions as reference data for reinforcement learning (RL) and imitation learning on a humanoid robot.

Labels

Master Thesis

Description

Goal

Contact Details

More information

Open this project... 

Published since: 2025-02-25

Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization ETH Competence Center - ETH AI Center

Hosts Li Chenhao , Kaufmann Manuel , Li Chenhao , Li Chenhao , Kaufmann Manuel , Li Chenhao

Topics Engineering and Technology

Learning Agile Dodgeball Behaviors for Humanoid Robots

Agility and rapid decision-making are vital for humanoid robots to safely and effectively operate in dynamic, unstructured environments. In human contexts—whether in crowded spaces, industrial settings, or collaborative environments—robots must be capable of reacting to fast, unpredictable changes in their surroundings. This includes not only planned navigation around static obstacles but also rapid responses to dynamic threats such as falling objects, sudden human movements, or unexpected collisions. Developing such reactive capabilities in legged robots remains a significant challenge due to the complexity of real-time perception, decision-making under uncertainty, and balance control. Humanoid robots, with their human-like morphology, are uniquely positioned to navigate and interact with human-centered environments. However, achieving fast, dynamic responses—especially while maintaining postural stability—requires advanced control strategies that integrate perception, motion planning, and balance control within tight time constraints. The task of dodging fast-moving objects, such as balls, provides an ideal testbed for studying these capabilities. It encapsulates several core challenges: rapid object detection and trajectory prediction, real-time motion planning, dynamic stability maintenance, and reactive behavior under uncertainty. Moreover, it presents a simplified yet rich framework to investigate more general collision avoidance strategies that could later be extended to complex real-world interactions. In robotics, reactive motion planning for dynamic environments has been widely studied, but primarily in the context of wheeled robots or static obstacle fields. Classical approaches focus on precomputed motion plans or simple reactive strategies, often unsuitable for highly dynamic scenarios where split-second decisions are critical. In the domain of legged robotics, maintaining balance while executing rapid, evasive maneuvers remains a challenging problem. Previous work on dynamic locomotion has addressed agile behaviors like running, jumping, or turning (e.g., Hutter et al., 2016; Kim et al., 2019), but these movements are often planned in advance rather than triggered reactively. More recent efforts have leveraged reinforcement learning (RL) to enable robots to adapt to dynamic environments, demonstrating success in tasks such as obstacle avoidance, perturbation recovery, and agile locomotion (Peng et al., 2017; Hwangbo et al., 2019). However, many of these approaches still struggle with real-time constraints and robustness in high-speed, unpredictable scenarios. Perception-driven control in humanoids, particularly for tasks requiring fast reactions, has seen advances through sensor fusion, visual servoing, and predictive modeling. For example, integrating vision-based object tracking with dynamic motion planning has enabled robots to perform tasks like ball catching or blocking (Ishiguro et al., 2002; Behnke, 2004). Yet, dodging requires a fundamentally different approach: instead of converging toward an object (as in catching), the robot must predict and strategically avoid the object’s trajectory while maintaining balance—often in the presence of limited maneuvering time. Dodgeball-inspired robotics research has been explored in limited contexts, primarily using wheeled robots or simplified agents in simulations. Few studies have addressed the challenges of high-speed evasion combined with the complexities of humanoid balance and multi-joint coordination. This project aims to bridge that gap by developing learning-based methods that enable humanoid robots to reactively avoid fast-approaching objects in real time, while preserving stability and agility.

Labels

Master Thesis

Description

Goal

Contact Details

More information

Open this project... 

Published since: 2025-02-25

Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization ETH Competence Center - ETH AI Center

Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao

Topics Engineering and Technology

Learning Real-time Human Motion Tracking on a Humanoid Robot

Humanoid robots, designed to mimic the structure and behavior of humans, have seen significant advancements in kinematics, dynamics, and control systems. Teleoperation of humanoid robots involves complex control strategies to manage bipedal locomotion, balance, and interaction with environments. Research in this area has focused on developing robots that can perform tasks in environments designed for humans, from simple object manipulation to navigating complex terrains. Reinforcement learning has emerged as a powerful method for enabling robots to learn from interactions with their environment, improving their performance over time without explicit programming for every possible scenario. In the context of humanoid robotics and teleoperation, RL can be used to optimize control policies, adapt to new tasks, and improve the efficiency and safety of human-robot interactions. Key challenges include the high dimensionality of the action space, the need for safe exploration, and the transfer of learned skills across different tasks and environments. Integrating human motion tracking with reinforcement learning on humanoid robots represents a cutting-edge area of research. This approach involves using human motion data as input to train RL models, enabling the robot to learn more natural and human-like movements. The goal is to develop systems that can not only replicate human actions in real-time but also adapt and improve their responses over time through learning. Challenges in this area include ensuring real-time performance, dealing with the variability of human motion, and maintaining stability and safety of the humanoid robot.

Keywords

real-time, humanoid, reinforcement learning, representation learning

Labels

Master Thesis

Description

Contact Details

More information

Open this project... 

Published since: 2025-02-25

Organization ETH Competence Center - ETH AI Center

Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao

Topics Information, Computing and Communication Sciences

Loosely Guided Reinforcement Learning for Humanoid Parkour

Humanoid robots hold the promise of navigating complex, human-centric environments with agility and adaptability. However, training these robots to perform dynamic behaviors such as parkour—jumping, climbing, and traversing obstacles—remains a significant challenge due to the high-dimensional state and action spaces involved. Traditional Reinforcement Learning (RL) struggles in such settings, primarily due to sparse rewards and the extensive exploration needed for complex tasks. This project proposes a novel approach to address these challenges by incorporating loosely guided references into the RL process. Instead of relying solely on task-specific rewards or complex reward shaping, we introduce a simplified reference trajectory that serves as a guide during training. This trajectory, often limited to the robot's base movement, reduces the exploration burden without constraining the policy to strict tracking, allowing the emergence of diverse and adaptable behaviors. Reinforcement Learning has demonstrated remarkable success in training agents for tasks ranging from game playing to robotic manipulation. However, its application to high-dimensional, dynamic tasks like humanoid parkour is hindered by two primary challenges: Exploration Complexity: The vast state-action space of humanoids leads to slow convergence, often requiring millions of training steps. Reward Design: Sparse rewards make it difficult for the agent to discover meaningful behaviors, while dense rewards demand intricate and often brittle design efforts. By introducing a loosely guided reference—a simple trajectory representing the desired flow of the task—we aim to reduce the exploration space while maintaining the flexibility of RL. This approach bridges the gap between pure RL and demonstration-based methods, enabling the learning of complex maneuvers like climbing, jumping, and dynamic obstacle traversal without heavy reliance on reward engineering or exact demonstrations.

Keywords

humanoid, reinforcement learning, loosely guided

Labels

Master Thesis

Description

Contact Details

More information

Open this project... 

Published since: 2025-02-25

Organization ETH Competence Center - ETH AI Center

Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao

Topics Information, Computing and Communication Sciences

Learning World Models for Legged Locomotion

Model-based reinforcement learning learns a world model from which an optimal control policy can be extracted. Understanding and predicting the forward dynamics of legged systems is crucial for effective control and planning. Forward dynamics involve predicting the next state of the robot given its current state and the applied actions. While traditional physics-based models can provide a baseline understanding, they often struggle with the complexities and non-linearities inherent in real-world scenarios, particularly due to the varying contact patterns of the robot's feet with the ground. The project aims to develop and evaluate neural network-based models for predicting the dynamics of legged environments, focusing on accounting for varying contact patterns and non-linearities. This involves collecting and preprocessing data from various simulation environment experiments, designing neural network architectures that incorporate necessary structures, and exploring hybrid models that combine physics-based predictions with neural network corrections. The models will be trained and evaluated on prediction autoregressive accuracy, with an emphasis on robustness and generalization capabilities across different noise perturbations. By the end of the project, the goal is to achieve an accurate, robust, and generalizable predictive model for the forward dynamics of legged systems.

Keywords

forward dynamics, non-smooth dynamics, neural networks, model-based reinforcement learning

Labels

Master Thesis

Description

Contact Details

More information

Open this project... 

Published since: 2025-02-25

Organization Robotic Systems Lab

Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao

Topics Engineering and Technology

Continual Learning and Domain Adaptation Techniques for a Waste Monitoring System on an Ocean Cleanup Vessel

The Autonomous River Cleanup (ARC) is a student-led initiative supported by the Robotic Systems Lab, focused on tackling riverine waste pollution. In partnership with The SeaCleaners, a Swiss NGO, this thesis aims to develop a self-improving onboard waste quantification system for the “Mobula 10” vessel collecting floating waste in the South East Asian Sea. Currently, waste quantification relies on manually counting collected items. The goal of this thesis is to automate the process using computer vision and hardware solutions tailored to the vessel’s infrastructure and the environmental conditions on the sea. Key to this effort will be the integration of continual learning [1] and domain adaptation [2] techniques for computer vision algorithms to adapt models to diverse and changing waste items, ensuring consistent performance without full retraining. Lastly, the system will be evaluated in real-world conditions to propose further improvements.

Keywords

Computer Vision, Continual Learning, Field Testing

Labels

Master Thesis

Contact Details

More information

Open this project... 

Published since: 2025-02-19 , Earliest start: 2025-02-19 , Latest end: 2025-12-31

Organization Robotic Systems Lab

Hosts Stolle Jonas

Topics Engineering and Technology

Multi-View Detection and Classification under Occlusions

The Autonomous River Cleanup (ARC) is a student-led initiative supported by the Robotic Systems Lab tackling the problem of riverine waste. By joining ARC, you will help improve the vision pipeline in our robotic sorting station. Currently, we first detect and classify items in a detection box without occlusion of the conveyor belt and re-detect them in the robot workspace by performing object tracking to bridge the gap. This approach has proven to be computationally expensive and requires extensive engineering to handle occlusions. Instead, we aim to use multiple cameras pointed at the workspace of the robotic arm to perform occlusion-robust detection and classification of waste objects. By combining the information from the two cameras in an end-to-end model, we aim to obtain higher confidence detections for items visible by both cameras and detections of partially occluded items only visible by one camera.

Keywords

Object Detection & Classification, Computer Vision

Labels

Semester Project

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-02-19 , Earliest start: 2025-02-19 , Latest end: 2025-06-30

Organization Robotic Systems Lab

Hosts Stolle Jonas

Topics Engineering and Technology

Supervised learning for loco-manipulation

To spot arm operations, we propose a multi-phase approach combining supervised learning and reinforcement learning (RL). First, we will employ supervised learning to develop a model for solving inverse kinematics (IK), enabling precise joint angle calculations from desired end-effector pose. Next, we will utilize another supervised learning technique to build a collision avoidance model, trained to predict and avoid self-collisions based on arm configurations and environmental data. With these pre-trained networks, we will then integrate RL to generate dynamic and safe arm-motion plans. The RL agent will leverage the IK and collision avoidance models to optimize arm trajectories, ensuring efficient and collision-free movements. This entire pipeline could be back propagated while promising to enhance the accuracy, safety, and flexibility of robotic arm operations in complex environments.

Keywords

Spot, Supervised learning, loco-manipulation

Labels

Master Thesis , ETH Zurich (ETHZ)

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-02-10 , Earliest start: 2025-02-10 , Latest end: 2026-03-01

Organization Robotic Systems Lab

Hosts Mirrazavi Sina

Topics Information, Computing and Communication Sciences

Model-Based Reinforcement Learning for Loco-manipulation

This project aims to develop a model-based reinforcement learning (RL) framework to enable quadruped robots to perform dynamic locomotion and manipulation simultaneously by leveraging advanced model-based RL algorithms such as DeamerV3, TDMPC2 and SAM-RL. We will develop control policies that can predict future states and rewards, enabling the robot to adapt its behavior on-the-fly. The primary focus will be on achieving stable and adaptive walking patterns while reaching and grasping objects. The outcome will provide insights into the integration of complex behaviors in robotic systems, with potential applications in service robotics and automated object handling.

Labels

Master Thesis , ETH Zurich (ETHZ)

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-02-10 , Earliest start: 2025-02-10 , Latest end: 2026-02-10

Organization Robotic Systems Lab

Hosts Mirrazavi Sina

Topics Information, Computing and Communication Sciences

Integrating OpenVLA for Vision-Language-Driven Loco-Manipulation robotics scenarios

This thesis proposes to integrate and adapt the OpenVLA (Open-Source Vision-Language-Action) model to control the Spot robotic arm for performing complex grasping and placing tasks. The study will focus on enabling the robot to recognize, grasp, and organize various toy-sized kitchen items based on human instructions. By leveraging OpenVLA's robust multimodal capabilities, this project aims to bridge the gap between human intent and robotic actions, enabling seamless task execution in unstructured environments. The research will explore the feasibility of fine-tuning OpenVLA for task-specific operations and evaluate its performance in real-world scenarios, providing valuable insights for advancing multimodal robotics.

Labels

Master Thesis , ETH Zurich (ETHZ)

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-02-10 , Earliest start: 2025-02-10 , Latest end: 2026-02-10

Organization Robotic Systems Lab

Hosts Mirrazavi Sina

Topics Information, Computing and Communication Sciences

Differentiable Simulation for Precise End-Effector Tracking

Unlock the potential of differentiable simulation on ALMA, a quadrupedal robot equipped with a robotic arm. Differentiable simulation enables precise gradient-based optimization, promising greater tracking accuracy and efficiency compared to standard reinforcement learning approaches. This project dives into advanced simulation and control techniques, paving the way for improvements in robotic trajectory tracking.

Keywords

Differentiable Simulation, Learning, ALMA

Labels

Semester Project , Bachelor Thesis , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-02-07 , Earliest start: 2025-01-27

Organization Robotic Systems Lab

Hosts Mittal Mayank , Schwarke Clemens , Klemm Victor

Topics Information, Computing and Communication Sciences

Modeling and Simulation for Earthwork in Digital Twin

In this work, we aim to build a digital twin of our autonomous hydraulic excavator, leveraging Mathworks technology for high-fidelity modeling. This will be used in the future to test and benchmark our learning-based controllers.

Keywords

Modeling, Hydraulics, Excavation, Industry

Labels

Semester Project , Master Thesis , ETH Zurich (ETHZ)

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-02-06 , Earliest start: 2025-03-03

Organization Robotic Systems Lab

Hosts Spinelli Filippo , Nan Fang

Topics Information, Computing and Communication Sciences , Engineering and Technology

Reinforcement Learning for Excavation Planning In Terra

We aim to develop a reinforcement learning-based global excavation planner that can plan for the long term and execute a wide range of excavation geometries. The system will be deployed on our legged excavator.

Keywords

Keywords: Reinforcement learning, task planning

Labels

Semester Project , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-02-03 , Earliest start: 2025-03-01 , Latest end: 2025-08-31

Organization Robotic Systems Lab

Hosts Terenzi Lorenzo

Topics Information, Computing and Communication Sciences

Model Based Reinforcement Learning

We want to train an excavator agent to learn in a variety of soil using a fast, GPU-accelerated soil particle simulator in Isaac Sim.

Keywords

particle simulation, omniverse, warp, reinforcement learning, model based reinforcement learning.

Labels

Semester Project , Master Thesis

Description

Work Packages

Requirements

More information

Open this project... 

Published since: 2025-02-03 , Earliest start: 2025-02-28 , Latest end: 2025-08-31

Organization Robotic Systems Lab

Hosts Egli Pascal Arturo , Terenzi Lorenzo

Topics Information, Computing and Communication Sciences , Engineering and Technology

Reinforcement Learning for Particle-Based Excavation in Isaac Sim

We want to train RL agents on our new particle simulator, accelerated on the GPU via warp in Isaac sim.

Keywords

particle simulation, omniverse, warp, reinforcement learning

Labels

Semester Project , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-02-03 , Earliest start: 2025-03-01 , Latest end: 2025-09-30

Organization Robotic Systems Lab

Hosts Egli Pascal Arturo , Mittal Mayank , Terenzi Lorenzo

Topics Information, Computing and Communication Sciences

Perceptive Reinforcement Learning for Exavation

In this project, our goal is to leverage precomputed embeddings(VAE in Isaacsim) from 3D earthworks scene reconstructions to train reinforcement learning agents. These embeddings, derived from incomplete point cloud data and reconstructed using an encoder-decoder neural network, will serve as latent representations. The main emphasis is on utilizing these representations to develop and train reinforcement learning policies for digging tasks.

Keywords

LIDAR, 3D reconstruction, Isaac gym, deep learning, perception, reinforcement learning

Labels

Semester Project , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-02-03 , Earliest start: 2025-01-01 , Latest end: 2025-08-31

Organization Robotic Systems Lab

Hosts Höller David , Terenzi Lorenzo

Topics Information, Computing and Communication Sciences

Reiforcement Learning of Pretrained Trasformer Models

We want to train RL agents on our new particle simulator, accelerated on the GPU via warp in Isaac sim.

Keywords

Keywords: particle simulation, omniverse, warp, reinforcement learning

Labels

Semester Project , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-02-03 , Earliest start: 2025-03-01 , Latest end: 2025-08-31

Organization Robotic Systems Lab

Hosts Terenzi Lorenzo

Topics Information, Computing and Communication Sciences

Multiagent Reinforcement Learning in Terra

We want to train multiple agents in the Terra environment, a fully end-to-end GPU-accelerated environment for RL training.

Keywords

multiagent reinforcement learning, jax, deep learning, planning

Labels

Semester Project , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-02-03 , Earliest start: 2025-03-01 , Latest end: 2025-07-31

Organization Robotic Systems Lab

Hosts Terenzi Lorenzo

Topics Information, Computing and Communication Sciences

Propose Your Own Robotics Challenge

This project invites you to step into the role of an innovator, encouraging you to identify challenges you are passionate about within the field of robotics. Rather than working on predefined problems, you will have the freedom to propose your own project ideas, address real-world issues, or explore cutting-edge topics. This project allows you to define your own research journey.

Keywords

Robotics, Research

Labels

Semester Project , Bachelor Thesis , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-01-28 , Earliest start: 2025-01-27

Organization Robotic Systems Lab

Hosts Schwarke Clemens , Bjelonic Filip , Klemm Victor

Topics Information, Computing and Communication Sciences

Data Driven Simulation for End-to-End Navigation

Investigate how neural rendering can become the backbone of comprehensive, next generation data-driven simulation

Keywords

Neural rendering, Simulation

Labels

Internship , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-01-24 , Earliest start: 2025-01-27

Organization Robotic Systems Lab

Hosts Kneip Laurent

Topics Information, Computing and Communication Sciences , Engineering and Technology

Evolving Minds: Neuroevolution for Legged Locomotion

This project explores the use of neuroevolution for optimizing control policies in legged robots, moving away from classical gradient-based methods like PPO. Neuroevolution directly optimizes network parameters and structures, potentially offering advantages in environments with sparse rewards, while requiring fewer hyperparameters to tune. By leveraging genetic algorithms and evolutionary strategies, the project aims to develop efficient controllers for complex locomotion tasks. With computational capabilities doubling approximately every two years as predicted by Moore's Law, neuroevolution offers a promising approach for scaling intelligent control systems.

Keywords

Evolutionary Algorithms, Reinforcement Learning, Quadrupeds, Legged Locomotion

Labels

Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-01-22 , Earliest start: 2025-01-27

Organization Robotic Systems Lab

Hosts Bjelonic Filip , Schwarke Clemens

Topics Information, Computing and Communication Sciences

Design of a Compliant Mechanism for Human-Robot Collaborative Transportation with Non-Holonomic Robots

Human-robot collaboration is an attractive option in many industries for transporting long and heavy items with a single operator. In this project, we aim to enable HRC transportation with a non-holonomic robotic base platform by designing a compliant manipulation mechanism, inspired by systems like the Omnid Mocobots.

Keywords

Human-robot collaboration Collaborative transportation Non-holonomic robot Mobile manipulation

Labels

Master Thesis , ETH Zurich (ETHZ)

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2025-01-16 , Earliest start: 2024-07-08

Applications limited to ETH Zurich

Organization Robotic Systems Lab

Hosts Kindle Julien , Bray Francesca

Topics Information, Computing and Communication Sciences , Engineering and Technology

How to Touch: Exploring Tactile Representations for Reinforcement Learning

Developing and benchmarking tactile representations for dexterous manipulation tasks using reinforcement learning.

Keywords

Reinforcement Learning, Dexterous Manipulation, Tactile Sensing

Labels

Semester Project , Bachelor Thesis , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2025-01-08 , Earliest start: 2024-12-15 , Latest end: 2025-06-01

Applications limited to ETH Zurich

Organization Robotic Systems Lab

Hosts Bhardwaj Arjun , Zurbrügg René

Topics Information, Computing and Communication Sciences

BEV meets Semantic traversability

Enable Birds-Eye-View perception on autonomous mobile robots for human-like navigation.

Keywords

Semantic Traversability, Birds-Eye-View, Localization, SLAM, Object Detection

Labels

Master Thesis , ETH Zurich (ETHZ)

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2024-12-18 , Earliest start: 2025-01-15 , Latest end: 2025-10-31

Organization Robotic Systems Lab

Hosts Gawel Abel

Topics Information, Computing and Communication Sciences , Engineering and Technology

Scene graphs for robot navigation and reasoning

Elevate semantic scene graphs to a new level and perform semantically-guided navigation and interaction with real robots at The AI Institute.

Keywords

Scene graphs, SLAM, Navigation, Spacial Reasoning, 3D reconstruction, Semantics

Labels

Master Thesis , ETH Zurich (ETHZ)

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2024-12-18 , Earliest start: 2025-01-15 , Latest end: 2025-10-31

Organization Robotic Systems Lab

Hosts Gawel Abel

Topics Information, Computing and Communication Sciences , Engineering and Technology

Digital Twin for Spot's Home

MOTIVATION ⇾ Creating a digital twin of the robot's environment is crucial for several reasons: 1. Simulate Different Robots: Test various robots in a virtual environment, saving time and resources. 2. Accurate Evaluation: Precisely assess robot interactions and performance. 3. Enhanced Flexibility: Easily modify scenarios to develop robust systems. 4. Cost Efficiency: Reduce costs by identifying issues in virtual simulations. 5. Scalability: Replicate multiple environments for comprehensive testing. PROPOSAL We propose to create a digital twin of our Semantic environment, designed in your preferred graphics Platform to be able to simulate Reinforcement Learning agents in the digital environment, to create a unified evaluation platform for robotic tasks.

Keywords

Digital Twin, Robotics

Labels

Semester Project , Master Thesis

Contact Details

More information

Open this project... 

Published since: 2024-12-17 , Earliest start: 2025-01-05

Applications limited to University of Zurich , ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization Computer Vision and Geometry Group

Hosts Blum Hermann , Portela Tifanny , Bauer Zuria, Dr. , Trisovic Jelena

Topics Information, Computing and Communication Sciences

KALLAX Benchmark: Evaluating Household Tasks

Motivation ⇾ There are three ways to evaluate robots for pick-and-place tasks at home: 1. Simulation setups: High reproducibility but hard to simulate real-world complexities and perception noise. 2. Competitions: Good for comparing overall systems but require significant effort and can't be done frequently. 3. Custom lab setups: Common but lead to overfitting and lack comparability between labs. Proposal ⇾ We propose using IKEA furniture to create standardized, randomized setups that researchers can easily replicate. E.g, a 4x4 KALLAX unit with varying door knobs and drawer positions, generating tasks like "move the cup from the upper right shelf into the black drawer." This prevents overfitting and allows for consistent evaluation across different labs.

Keywords

Benchmakr, Robotics, pick-and-place

Labels

Semester Project , Master Thesis

Contact Details

More information

Open this project... 

Published since: 2024-12-17 , Earliest start: 2025-01-06

Applications limited to University of Zurich , ETH Zurich , Swiss National Science Foundation , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization Computer Vision and Geometry Group

Hosts Blum Hermann , Bauer Zuria, Dr. , Zurbrügg René

Topics Information, Computing and Communication Sciences

Adapting to Injuries for Dexterous In-Hand Manipulation

Develop a reinforcement learning-based method for training adaptive policies for dexterous in-hand manipulation systems to deal with actuator failure on the fly.

Keywords

Dexterous Manipulation, Reinforcement Learning, Adaptive Learning

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2024-12-12 , Earliest start: 2024-12-16 , Latest end: 2025-06-01

Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization Robotic Systems Lab

Hosts Bhardwaj Arjun , Ma Yuntao

Topics Information, Computing and Communication Sciences

Extending Functional Scene Graphs to Include Articulated Object States

While traditional [1] and functional [2] scene graphs are capable of capturing the spatial relationships and functional interactions between objects and spaces, they encode each object as static, with fixed geometry. In this project, we aim to enable the estimation of the state of articulated objects and include it in the functional scene graph.

Keywords

scene understanding, scene graph, exploration

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2024-12-11

Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization Computer Vision and Geometry Group

Hosts Bauer Zuria, Dr. , Trisovic Jelena , Zurbrügg René

Topics Information, Computing and Communication Sciences , Engineering and Technology

Active Object Localization with Touch

Develop active exploration strategies for object identification and localization with tactile feedback.

Keywords

Dexterous Manipulation, Object Retrieval, Active Localization, Tactile Sensing

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2024-12-11 , Earliest start: 2024-12-15 , Latest end: 2025-06-01

Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization Robotic Systems Lab

Hosts Bhardwaj Arjun , Zurbrügg René

Topics Information, Computing and Communication Sciences

Learning-based object orientation prediction for handovers

Humans are exceptional at handovers. Besides timing and spatial precision, they also have a high-level understanding of how the other person wants to use the object that is handed over. This information is needed to hand over an object, such that it can be used directly for a specific task. While robots can reason about grasp affordances, the integration of this information with perception and control is missing.

Keywords

Robot-Human Handover, Human-Robot-Interaction, Mobile Manipulation, Robotics

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2024-12-10 , Earliest start: 2025-02-01 , Latest end: 2025-12-31

Organization Robotic Systems Lab

Hosts Scheidemann Carmen , Tulbure Andreea

Topics Information, Computing and Communication Sciences , Engineering and Technology

Evolve to Grasp: Learning Optimal Finger Configuration for a Dexterous Multifingered Hand

Use evolutionary algorithms with analytical force closure metrics to learn the optimal morphology of a dexterous hand.

Keywords

Evolutionary Algorithm, Machine Learning, Grasping, Robotics

Labels

Semester Project , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2024-12-09 , Earliest start: 2024-12-09 , Latest end: 2025-10-31

Applications limited to ETH Zurich

Organization Robotic Systems Lab

Hosts Church Joseph , Zurbrügg René

Topics Information, Computing and Communication Sciences

Differential Particle Simulation for Robotics

This project focuses on applying differential particle-based simulation to address challenges in simulating real-world robotic tasks involving interactions with fluids, granular materials, and soft objects. Leveraging the differentiability of simulations, the project aims to enhance simulation accuracy with limited real-world data and explore learning robotic control using first-order gradient information.

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2024-12-09 , Earliest start: 2025-01-01 , Latest end: 2025-12-31

Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization Robotic Systems Lab

Hosts Nan Fang , Ma Hao

Topics Engineering and Technology

Conformal Prediction for Distribution Shift Detection in Online Learning

This project investigates the use of conformal prediction for detecting distribution shifts in online learning scenarios, with a focus on robotics applications. Distribution shifts, arising from deviations in task distributions or changes in robot dynamics, pose significant challenges to online learning systems by impacting learning efficiency and model performance. The project aims to develop a robust detection algorithm to address these shifts, classifying task distribution shifts as outliers while dynamically retraining models for characteristic shifts.

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2024-12-09 , Earliest start: 2025-01-01 , Latest end: 2025-12-31

Organization Robotic Systems Lab

Hosts Ma Hao , Nan Fang

Topics Information, Computing and Communication Sciences , Engineering and Technology

Visual Language Models for Long-Term Planning

This project uses Visual Language Models (VLMs) for high-level planning and supervision in construction tasks, enabling task prioritization, dynamic adaptation, and multi-robot collaboration for excavation and site management. prioritization, dynamic adaptation, and multi-robot collaboration for excavation and site management

Keywords

Visual Language Models, Long-term planning, Robotics

Labels

Semester Project , Master Thesis

Description

Work Packages

Contact Details

More information

Open this project... 

Published since: 2024-12-06 , Earliest start: 2025-01-24 , Latest end: 2025-10-29

Organization Robotic Systems Lab

Hosts Terenzi Lorenzo

Topics Information, Computing and Communication Sciences

Diffusion-based Shared Autonomy System for Telemanipulation

Robots may not be able to complete tasks fully autonomously in unstructured or unseen environments, however direct teleoperation from human operators may also be challenging due to the difficulty of providing full situational awareness to the operator as well as degradation in communication leading to the loss of control authority. This motivates the use of shared autonomy for assisting the operator thereby enhancing the performance during the task. In this project, we aim to develop a shared autonomy framework for teleoperation of manipulator arms, to assist non-expert users or in the presence of degraded communication. Imitation learning, such as diffusion models, have emerged as a popular and scalable approach for learning manipulation tasks [1, 2]. Additionally, recent works have combined this with partial diffusion to enable shared autonomy [3]. However, the tasks were restricted to simple 2D domains. In this project, we wish to extend previous work in the lab using diffusion-based imitation learning, to enable shared autonomy for non-expert users to complete unseen tasks or in degraded communication environments.

Keywords

Imitation learning, Robotics, Manipulation, Teleoperation

Labels

Semester Project , ETH Zurich (ETHZ)

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2024-12-02 , Earliest start: 2024-11-01 , Latest end: 2025-11-01

Applications limited to ETH Zurich , University of Zurich

Organization Robotic Systems Lab

Hosts Elanjimattathil Aravind

Topics Information, Computing and Communication Sciences , Engineering and Technology

Lifelike Agility on ANYmal by Learning from Animals

The remarkable agility of animals, characterized by their rapid, fluid movements and precise interaction with their environment, serves as an inspiration for advancements in legged robotics. Recent progress in the field has underscored the potential of learning-based methods for robot control. These methods streamline the development process by optimizing control mechanisms directly from sensory inputs to actuator outputs, often employing deep reinforcement learning (RL) algorithms. By training in simulated environments, these algorithms can develop locomotion skills that are subsequently transferred to physical robots. Although this approach has led to significant achievements in achieving robust locomotion, mimicking the wide range of agile capabilities observed in animals remains a significant challenge. Traditionally, manually crafted controllers have succeeded in replicating complex behaviors, but their development is labor-intensive and demands a high level of expertise in each specific skill. Reinforcement learning offers a promising alternative by potentially reducing the manual labor involved in controller development. However, crafting learning objectives that lead to the desired behaviors in robots also requires considerable expertise, specific to each skill.

Keywords

learning from demonstrations, imitation learning, reinforcement learning

Labels

Master Thesis

Description

Contact Details

More information

Open this project... 

Published since: 2024-11-26

Organization ETH Competence Center - ETH AI Center

Hosts Li Chenhao , Li Chenhao , Klemm Victor

Topics Information, Computing and Communication Sciences

Pushing the Limit of Quadruped Running Speed with Autonomous Curriculum Learning

The project aims to explore curriculum learning techniques to push the limits of quadruped running speed using reinforcement learning. By systematically designing and implementing curricula that guide the learning process, the project seeks to develop a quadruped controller capable of achieving the fastest possible forward locomotion. This involves not only optimizing the learning process but also ensuring the robustness and adaptability of the learned policies across various running conditions.

Keywords

curriculum learning, fast locomotion

Labels

Master Thesis

Description

Contact Details

More information

Open this project... 

Published since: 2024-11-26

Organization Robotic Systems Lab

Hosts Li Chenhao , Bagatella Marco , Li Chenhao , Li Chenhao , Li Chenhao

Topics Engineering and Technology

Humanoid Locomotion Learning and Finetuning from Human Feedback

In the burgeoning field of deep reinforcement learning (RL), agents autonomously develop complex behaviors through a process of trial and error. Yet, the application of RL across various domains faces notable hurdles, particularly in devising appropriate reward functions. Traditional approaches often resort to sparse rewards for simplicity, though these prove inadequate for training efficient agents. Consequently, real-world applications may necessitate elaborate setups, such as employing accelerometers for door interaction detection, thermal imaging for action recognition, or motion capture systems for precise object tracking. Despite these advanced solutions, crafting an ideal reward function remains challenging due to the propensity of RL algorithms to exploit the reward system in unforeseen ways. Agents might fulfill objectives in unexpected manners, highlighting the complexity of encoding desired behaviors, like adherence to social norms, into a reward function. An alternative strategy, imitation learning, circumvents the intricacies of reward engineering by having the agent learn through the emulation of expert behavior. However, acquiring a sufficient number of high-quality demonstrations for this purpose is often impractically costly. Humans, in contrast, learn with remarkable autonomy, benefiting from intermittent guidance from educators who provide tailored feedback based on the learner's progress. This interactive learning model holds promise for artificial agents, offering a customized learning trajectory that mitigates reward exploitation without extensive reward function engineering. The challenge lies in ensuring the feedback process is both manageable for humans and rich enough to be effective. Despite its potential, the implementation of human-in-the-loop (HiL) RL remains limited in practice. Our research endeavors to significantly lessen the human labor involved in HiL learning, leveraging both unsupervised pre-training and preference-based learning to enhance agent development with minimal human intervention.

Keywords

reinforcement learning from human feedback, preference learning

Labels

Master Thesis

Description

Goal

Contact Details

More information

Open this project... 

Published since: 2024-11-26

Organization ETH Competence Center - ETH AI Center

Hosts Li Chenhao , Li Chenhao , Chen Xin , Li Chenhao

Topics Information, Computing and Communication Sciences , Engineering and Technology

Online Safe Locomotion Learning in the Wild

Reinforcement learning (RL) can potentially solve complex problems in a purely data-driven manner. Still, the state-of-the-art in applying RL in robotics, relies heavily on high-fidelity simulators. While learning in simulation allows to circumvent sample complexity challenges that are common in model-free RL, even slight distribution shift ("sim-to-real gap") between simulation and the real system can cause these algorithms to easily fail. Recent advances in model-based reinforcement learning have led to superior sample efficiency, enabling online learning without a simulator. Nonetheless, learning online cannot cause any damage and should adhere to safety requirements (for obvious reasons). The proposed project aims to demonstrate how existing safe model-based RL methods can be used to solve the foregoing challenges.

Keywords

safe mode-base RL, online learning, legged robotics

Labels

Master Thesis

Description

Contact Details

More information

Open this project... 

Published since: 2024-11-26

Organization ETH Competence Center - ETH AI Center

Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao

Topics Engineering and Technology

Autonomous Curriculum Learning for Increasingly Challenging Tasks

While the history of machine learning so far largely encompasses a series of problems posed by researchers and algorithms that learn their solutions, an important question is whether the problems themselves can be generated by the algorithm at the same time as they are being solved. Such a process would in effect build its own diverse and expanding curricula, and the solutions to problems at various stages would become stepping stones towards solving even more challenging problems later in the process. Consider the realm of legged locomotion: Training a robot via reinforcement learning to track a velocity command illustrates this concept. Initially, tracking a low velocity is simpler due to algorithm initialization and environmental setup. By manually crafting a curriculum, we can start with low-velocity targets and incrementally increase them as the robot demonstrates competence. This method works well when the difficulty correlates clearly with the target, as with higher velocities or more challenging terrains. However, challenges arise when the relationship between task difficulty and control parameters is unclear. For instance, if a parameter dictates various human dance styles for the robot to mimic, it's not obvious whether jazz is easier than hip-hop. In such scenarios, the difficulty distribution does not align with the control parameter. How, then, can we devise an effective curriculum? In the conventional RSL training setting for locomotion over challenging terrains, there is also a handcrafted learning schedule dictating increasingly hard terrain levels but unified with multiple different types. With a smart autonomous curriculum learning algorithm, are we able to overcome separate terrain types asynchronously and thus achieve overall better performance or higher data efficiency?

Keywords

curriculum learning, open-ended learning, self-evolution, progressive task solving

Labels

Master Thesis

Description

Contact Details

More information

Open this project... 

Published since: 2024-11-26

Organization Robotic Systems Lab

Hosts Li Chenhao , Li Chenhao , Li Chenhao , Bagatella Marco , Li Chenhao

Topics Engineering and Technology

Humanoid Locomotion Learning with Human Motion Priors

Humanoid robots, designed to replicate human structure and behavior, have made significant strides in kinematics, dynamics, and control systems. Research aims to develop robots capable of performing tasks in human-centric settings, from simple object manipulation to navigating complex terrains. Reinforcement learning (RL) has proven to be a powerful method for enabling robots to learn from their environment, enhancing their performance over time without explicit programming for every possible scenario. In the realm of humanoid robotics, RL is used to optimize control policies, adapt to new tasks, and improve the efficiency and safety of human-robot interactions. However, one of the primary challenges is the high dimensionality of the action space, where handcrafted reward functions fall short of generating natural, lifelike motions. Incorporating motion priors into the learning process of humanoid robots addresses these challenges effectively. Motion priors can significantly reduce the exploration space in RL, leading to faster convergence and reduced training time. They ensure that learned policies prioritize stability and safety, reducing the risk of unpredictable or hazardous actions. Additionally, motion priors guide the learning process towards more natural, human-like movements, improving the robot's ability to perform tasks intuitively and seamlessly in human environments. Therefore, motion priors are crucial for efficient, stable, and realistic humanoid locomotion learning, enabling robots to better navigate and interact with the world around them.

Keywords

motion priors, humanoid, reinforcement learning, representation learning

Labels

Master Thesis

Description

Contact Details

More information

Open this project... 

Published since: 2024-11-26

Organization ETH Competence Center - ETH AI Center

Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao

Topics Information, Computing and Communication Sciences

DigGPT: Large Language Models for Excavation Planning

Large language models (LLMs) have shown the first sparks of artificial general intelligence. We want to test if GPT 4.0 can solve excavation planning problems.

Keywords

GPT, Large Language Models, Robotics, Deep Learning, Reinforcement Learning

Labels

Semester Project , Master Thesis

Description

Work Packages

Requirements

Contact Details

More information

Open this project... 

Published since: 2024-11-21 , Earliest start: 2025-01-01 , Latest end: 2025-08-31

Organization Robotic Systems Lab

Hosts Terenzi Lorenzo

Topics Engineering and Technology

VR in Habitat 3.0

Motivation: Explore the newly improved Habitat 3.0 simulator with a special focus on the Virtual Reality Features. This project is meant to be an exploration task on the Habitat 3.0 simulator, exploring all the newly introduced features focusing specifically on the implementation of virtual reality tools for scene navigation. The idea is to extend these features to self created environments in Unreal Engine that build uppon Habitat

Labels

Semester Project

Contact Details

More information

Open this project... 

Published since: 2024-11-06 , Earliest start: 2024-01-08

Applications limited to University of Zurich , Swiss National Science Foundation , ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne

Organization Computer Vision and Geometry Group

Hosts Blum Hermann , Bauer Zuria, Dr. , Sun Boyang , Zurbrügg René

Topics Information, Computing and Communication Sciences

Generalist Excavator Transformer

We want to develop a generalist digging agent that is able to do multiple tasks, such as digging and moving loose soil, and/or control multiple excavators. We plan to use decision transformers, trained on offline data, to accomplish these tasks.

Keywords

Offline reinforcement learning, transformers, autonomous excavation

Labels

Semester Project , Master Thesis

PLEASE LOG IN TO SEE DESCRIPTION

More information

Open this project... 

Published since: 2024-10-11 , Earliest start: 2025-03-01 , Latest end: 2025-08-31

Organization Robotic Systems Lab

Hosts Werner Lennart , Egli Pascal Arturo , Terenzi Lorenzo , Nan Fang , Zhang Weixuan

Topics Information, Computing and Communication Sciences

Student Theses in Industry

We have a large number of industry partners who search for excellent students to conduct their student theses at the company or at ETH but in close collaboration with them (joint supervision by industry and ETH). 

Ammann Group (Switzerland)

The Ammann Group is a worldwide leader in the manufacture of mixing plants, machinery, and services in the construction industry, with core competence in road construction and landscaping as well as in the transport infrastructure.

We are collaborating with Ammann to automate construction equipment

Maxon (Switzerland)

Maxon develops and builds electric drive systems that are among the best in the world. Their drive systems can be found wherever extreme precision and the highest quality standards are indispensable – on Earth, and on Mars.

Shunk (Germany)

Legged Wheel Chair

Simulation, control and design of a robotized wheelchair able to cross difficult terrain

external page Simulation, control and design of a robotized wheelchair able to cross difficult terrain

This project aims at extending a dynamic simulation and locomotion controllers for a robotized wheelchair able to handle difficult terrains including stairs. This project will prepare the prototype phase coming next. 

.

Note on plagiarism

We would like to suggest every student, irrespective of the type of project (Bachelor, Semester, Master, ...), to make himself/herself familiar with ETH rules regarding plagiarism

JavaScript has been disabled in your browser