Student Projects
Currently, the following student projects are available. Please contact the responsible supervisor and apply with your CV and transcripts.
In case you have project ideas related to any of these projects, take the opportunity and propose your own project!
We also offer Master Thesis and or exchange semesters at reknown universities around the world. Please contact us in case of interest!
Studies on Mechatronics
We offer students also to conduct their Studies on Mechatronics at our lab. In general, we recommend to do the Studies on Mechatronics in combination with the Bachelor Thesis, either as prepartory work the semester beofre or as extended study in parallel. If you want to do it independently, yiou can find prroposed projets also in the list below. Please directly apply with corresponsponding supervisor.
Reinforcement learning to reject large disturbances for a robotic arm
Quasi-direct drive robot arms experience high deflection in the presence of large end effector forces. This leads to stability issues when these forces change rapidly, such as when dropping a heavy object. Existing works use reinforcement learning to achieve end effector tracking[1] as well as force control[2]. In this project, you will extend upon this by developing a controller capable of tracking end effector poses in the presence of sudden, large changes in end effector disturbance. You will first deploy your controller on a standalone robotic arm and then work towards deployment on our ANYmal-based bimanual mobile manipulator. [1] Martín-Martín et al., Variable Impedance Control in End-Effector Space: An Action Space for Reinforcement Learning in Contact-Rich Tasks (IROS 2019) [2] Portela et al., Learning Force Control for Legged Manipulation (ICRA 2024)
Keywords
reinforcement learning, robot arm, mobile manipulation, force disturbance
Labels
Semester Project , Bachelor Thesis
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-03-07 , Earliest start: 2025-03-10
Organization Robotic Systems Lab
Hosts Mittal Mayank , Fischer Oliver
Topics Engineering and Technology
Novel Winch Control for Robotic Climbing
While legged robots have demonstrated impressive locomotion performance in structured environments, challenges persist in navigating steep natural terrain and loose, granular soil. These challenges extend to extraterrestrial environments and are relevant to future lunar, martian, and asteroidal missions. In order to explore the most extreme terrains, a novel winch system has been developed for the ANYmal robot platform. The winch could potentially be used as a fail-safe device to prevent falls during unassisted traverses of steep terrain, as well as an added driven degree of freedom for assisted ascending and descending of terrain too steep for unassisted traversal. The goal of this project is to develop control policies that utilize this new hardware and enable further climbing robot research.
Keywords
Robot, Space, Climbing, Winch, Control
Labels
Semester Project , Master Thesis , ETH Zurich (ETHZ)
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-03-05 , Earliest start: 2024-10-07
Organization Robotic Systems Lab
Hosts Vogel Dylan
Topics Information, Computing and Communication Sciences , Engineering and Technology
Beyond Value Functions: Stable Robot Learning with Monte-Carlo GRPO
Robotics is dominated by on-policy reinforcement learning: the paradigm of training a robot controller by iteratively interacting with the environment and maximizing some objective. A crucial idea to make this work is the Advantage Function. On each policy update, algorithms typically sum up the gradient log probabilities of all actions taken in the robot simulation. The advantage function increases or decreases the probabilities of these taken actions by comparing their “goodness” versus a baseline. Current advantage estimation methods use a value function to aggregate robot experience and hence decrease variance. This improves sample efficiency at the cost of introducing some bias. Stably training large language models via reinforcement learning is well-known to be a challenging task. A line of recent work [1, 2] has used Group-Relative Policy Optimization (GRPO) to achieve this feat. In GRPO, a series of answers are generated for each query-answer pair. The advantage is calculated based on a given answer being better than the average answer to the query. In this formulation, no value function is required. Can we adapt GRPO towards robot learning? Value Functions are known to cause issues in training stability [3] and a result in biased advantage estimates [4]. We are in the age of GPU-accelerated RL [5], training policies by simulating thousands of robot instances simultaneously. This makes a new monte-carlo (MC) approach towards RL timely, feasible and appealing. In this project, the student will be tasked to investigate the limitations of value-function based advantage estimation. Using GRPO as a starting point, the student will then develop MC-based algorithms that use the GPU’s parallel simulation capabilities for stable RL training for unbiased variance reduction while maintaining a competitive wall-clock time.
Keywords
Robot Learning, Reinforcement Learning, Monte Carlo RL, GRPO, Advantage Estimation
Labels
Semester Project , Bachelor Thesis , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-03-05
Organization Robotic Systems Lab
Hosts Klemm Victor
Topics Information, Computing and Communication Sciences , Engineering and Technology , Behavioural and Cognitive Sciences
Manipulation beyond Single End-Effector
The goal of the project is to extend our prior works to make ANYmal with an arm use its different end-effectors for whole-body mobile manipulation.
Keywords
reinforcement learning, robotics, perception, robot learning
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-03-03 , Earliest start: 2025-03-17
Organization Robotic Systems Lab
Hosts Mittal Mayank
Topics Information, Computing and Communication Sciences , Behavioural and Cognitive Sciences
Volumetric Bucket-Fill Estimation
Gravis Robotics is an ETH spinoff from the Robotic Systems Lab (RSL) working on the automation of heavy machinery (https://gravisrobotics.com/). In this project, you will be working with the Gravis team to develop a perceptive bucket-fill estimation system. You will conduct your project at Gravis under joint supervision from RSL.
Keywords
Autonomous Excavation
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-02-28 , Earliest start: 2025-01-01 , Latest end: 2026-01-01
Organization Robotic Systems Lab
Hosts Egli Pascal Arturo
Topics Engineering and Technology
Leveraging Human Motion Data from Videos for Humanoid Robot Motion Learning
The advancement in humanoid robotics has reached a stage where mimicking complex human motions with high accuracy is crucial for tasks ranging from entertainment to human-robot interaction in dynamic environments. Traditional approaches in motion learning, particularly for humanoid robots, rely heavily on motion capture (MoCap) data. However, acquiring large amounts of high-quality MoCap data is both expensive and logistically challenging. In contrast, video footage of human activities, such as sports events or dance performances, is widely available and offers an abundant source of motion data. Building on recent advancements in extracting and utilizing human motion from videos, such as the method proposed in WHAM (refer to the paper "Learning Physically Simulated Tennis Skills from Broadcast Videos"), this project aims to develop a system that extracts human motion from videos and applies it to teach a humanoid robot how to perform similar actions. The primary focus will be on extracting dynamic and expressive motions from videos, such as soccer player celebrations, and using these extracted motions as reference data for reinforcement learning (RL) and imitation learning on a humanoid robot.
Labels
Master Thesis
Description
Goal
Contact Details
More information
Open this project... call_made
Published since: 2025-02-25
Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Kaufmann Manuel , Li Chenhao , Li Chenhao , Kaufmann Manuel , Li Chenhao
Topics Engineering and Technology
Learning Agile Dodgeball Behaviors for Humanoid Robots
Agility and rapid decision-making are vital for humanoid robots to safely and effectively operate in dynamic, unstructured environments. In human contexts—whether in crowded spaces, industrial settings, or collaborative environments—robots must be capable of reacting to fast, unpredictable changes in their surroundings. This includes not only planned navigation around static obstacles but also rapid responses to dynamic threats such as falling objects, sudden human movements, or unexpected collisions. Developing such reactive capabilities in legged robots remains a significant challenge due to the complexity of real-time perception, decision-making under uncertainty, and balance control. Humanoid robots, with their human-like morphology, are uniquely positioned to navigate and interact with human-centered environments. However, achieving fast, dynamic responses—especially while maintaining postural stability—requires advanced control strategies that integrate perception, motion planning, and balance control within tight time constraints. The task of dodging fast-moving objects, such as balls, provides an ideal testbed for studying these capabilities. It encapsulates several core challenges: rapid object detection and trajectory prediction, real-time motion planning, dynamic stability maintenance, and reactive behavior under uncertainty. Moreover, it presents a simplified yet rich framework to investigate more general collision avoidance strategies that could later be extended to complex real-world interactions. In robotics, reactive motion planning for dynamic environments has been widely studied, but primarily in the context of wheeled robots or static obstacle fields. Classical approaches focus on precomputed motion plans or simple reactive strategies, often unsuitable for highly dynamic scenarios where split-second decisions are critical. In the domain of legged robotics, maintaining balance while executing rapid, evasive maneuvers remains a challenging problem. Previous work on dynamic locomotion has addressed agile behaviors like running, jumping, or turning (e.g., Hutter et al., 2016; Kim et al., 2019), but these movements are often planned in advance rather than triggered reactively. More recent efforts have leveraged reinforcement learning (RL) to enable robots to adapt to dynamic environments, demonstrating success in tasks such as obstacle avoidance, perturbation recovery, and agile locomotion (Peng et al., 2017; Hwangbo et al., 2019). However, many of these approaches still struggle with real-time constraints and robustness in high-speed, unpredictable scenarios. Perception-driven control in humanoids, particularly for tasks requiring fast reactions, has seen advances through sensor fusion, visual servoing, and predictive modeling. For example, integrating vision-based object tracking with dynamic motion planning has enabled robots to perform tasks like ball catching or blocking (Ishiguro et al., 2002; Behnke, 2004). Yet, dodging requires a fundamentally different approach: instead of converging toward an object (as in catching), the robot must predict and strategically avoid the object’s trajectory while maintaining balance—often in the presence of limited maneuvering time. Dodgeball-inspired robotics research has been explored in limited contexts, primarily using wheeled robots or simplified agents in simulations. Few studies have addressed the challenges of high-speed evasion combined with the complexities of humanoid balance and multi-joint coordination. This project aims to bridge that gap by developing learning-based methods that enable humanoid robots to reactively avoid fast-approaching objects in real time, while preserving stability and agility.
Labels
Master Thesis
Description
Goal
Contact Details
More information
Open this project... call_made
Published since: 2025-02-25
Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao
Topics Engineering and Technology
Learning Real-time Human Motion Tracking on a Humanoid Robot
Humanoid robots, designed to mimic the structure and behavior of humans, have seen significant advancements in kinematics, dynamics, and control systems. Teleoperation of humanoid robots involves complex control strategies to manage bipedal locomotion, balance, and interaction with environments. Research in this area has focused on developing robots that can perform tasks in environments designed for humans, from simple object manipulation to navigating complex terrains. Reinforcement learning has emerged as a powerful method for enabling robots to learn from interactions with their environment, improving their performance over time without explicit programming for every possible scenario. In the context of humanoid robotics and teleoperation, RL can be used to optimize control policies, adapt to new tasks, and improve the efficiency and safety of human-robot interactions. Key challenges include the high dimensionality of the action space, the need for safe exploration, and the transfer of learned skills across different tasks and environments. Integrating human motion tracking with reinforcement learning on humanoid robots represents a cutting-edge area of research. This approach involves using human motion data as input to train RL models, enabling the robot to learn more natural and human-like movements. The goal is to develop systems that can not only replicate human actions in real-time but also adapt and improve their responses over time through learning. Challenges in this area include ensuring real-time performance, dealing with the variability of human motion, and maintaining stability and safety of the humanoid robot.
Keywords
real-time, humanoid, reinforcement learning, representation learning
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2025-02-25
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao
Topics Information, Computing and Communication Sciences
Loosely Guided Reinforcement Learning for Humanoid Parkour
Humanoid robots hold the promise of navigating complex, human-centric environments with agility and adaptability. However, training these robots to perform dynamic behaviors such as parkour—jumping, climbing, and traversing obstacles—remains a significant challenge due to the high-dimensional state and action spaces involved. Traditional Reinforcement Learning (RL) struggles in such settings, primarily due to sparse rewards and the extensive exploration needed for complex tasks. This project proposes a novel approach to address these challenges by incorporating loosely guided references into the RL process. Instead of relying solely on task-specific rewards or complex reward shaping, we introduce a simplified reference trajectory that serves as a guide during training. This trajectory, often limited to the robot's base movement, reduces the exploration burden without constraining the policy to strict tracking, allowing the emergence of diverse and adaptable behaviors. Reinforcement Learning has demonstrated remarkable success in training agents for tasks ranging from game playing to robotic manipulation. However, its application to high-dimensional, dynamic tasks like humanoid parkour is hindered by two primary challenges: Exploration Complexity: The vast state-action space of humanoids leads to slow convergence, often requiring millions of training steps. Reward Design: Sparse rewards make it difficult for the agent to discover meaningful behaviors, while dense rewards demand intricate and often brittle design efforts. By introducing a loosely guided reference—a simple trajectory representing the desired flow of the task—we aim to reduce the exploration space while maintaining the flexibility of RL. This approach bridges the gap between pure RL and demonstration-based methods, enabling the learning of complex maneuvers like climbing, jumping, and dynamic obstacle traversal without heavy reliance on reward engineering or exact demonstrations.
Keywords
humanoid, reinforcement learning, loosely guided
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2025-02-25
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao
Topics Information, Computing and Communication Sciences
Learning World Models for Legged Locomotion
Model-based reinforcement learning learns a world model from which an optimal control policy can be extracted. Understanding and predicting the forward dynamics of legged systems is crucial for effective control and planning. Forward dynamics involve predicting the next state of the robot given its current state and the applied actions. While traditional physics-based models can provide a baseline understanding, they often struggle with the complexities and non-linearities inherent in real-world scenarios, particularly due to the varying contact patterns of the robot's feet with the ground. The project aims to develop and evaluate neural network-based models for predicting the dynamics of legged environments, focusing on accounting for varying contact patterns and non-linearities. This involves collecting and preprocessing data from various simulation environment experiments, designing neural network architectures that incorporate necessary structures, and exploring hybrid models that combine physics-based predictions with neural network corrections. The models will be trained and evaluated on prediction autoregressive accuracy, with an emphasis on robustness and generalization capabilities across different noise perturbations. By the end of the project, the goal is to achieve an accurate, robust, and generalizable predictive model for the forward dynamics of legged systems.
Keywords
forward dynamics, non-smooth dynamics, neural networks, model-based reinforcement learning
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2025-02-25
Organization Robotic Systems Lab
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao
Topics Engineering and Technology
Continual Learning and Domain Adaptation Techniques for a Waste Monitoring System on an Ocean Cleanup Vessel
The Autonomous River Cleanup (ARC) is a student-led initiative supported by the Robotic Systems Lab, focused on tackling riverine waste pollution. In partnership with The SeaCleaners, a Swiss NGO, this thesis aims to develop a self-improving onboard waste quantification system for the “Mobula 10” vessel collecting floating waste in the South East Asian Sea. Currently, waste quantification relies on manually counting collected items. The goal of this thesis is to automate the process using computer vision and hardware solutions tailored to the vessel’s infrastructure and the environmental conditions on the sea. Key to this effort will be the integration of continual learning [1] and domain adaptation [2] techniques for computer vision algorithms to adapt models to diverse and changing waste items, ensuring consistent performance without full retraining. Lastly, the system will be evaluated in real-world conditions to propose further improvements.
Keywords
Computer Vision, Continual Learning, Field Testing
Labels
Master Thesis
Contact Details
More information
Open this project... call_made
Published since: 2025-02-19 , Earliest start: 2025-02-19 , Latest end: 2025-12-31
Organization Robotic Systems Lab
Hosts Stolle Jonas
Topics Engineering and Technology
Multi-View Detection and Classification under Occlusions
The Autonomous River Cleanup (ARC) is a student-led initiative supported by the Robotic Systems Lab tackling the problem of riverine waste. By joining ARC, you will help improve the vision pipeline in our robotic sorting station. Currently, we first detect and classify items in a detection box without occlusion of the conveyor belt and re-detect them in the robot workspace by performing object tracking to bridge the gap. This approach has proven to be computationally expensive and requires extensive engineering to handle occlusions. Instead, we aim to use multiple cameras pointed at the workspace of the robotic arm to perform occlusion-robust detection and classification of waste objects. By combining the information from the two cameras in an end-to-end model, we aim to obtain higher confidence detections for items visible by both cameras and detections of partially occluded items only visible by one camera.
Keywords
Object Detection & Classification, Computer Vision
Labels
Semester Project
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-02-19 , Earliest start: 2025-02-19 , Latest end: 2025-06-30
Organization Robotic Systems Lab
Hosts Stolle Jonas
Topics Engineering and Technology
Supervised learning for loco-manipulation
To spot arm operations, we propose a multi-phase approach combining supervised learning and reinforcement learning (RL). First, we will employ supervised learning to develop a model for solving inverse kinematics (IK), enabling precise joint angle calculations from desired end-effector pose. Next, we will utilize another supervised learning technique to build a collision avoidance model, trained to predict and avoid self-collisions based on arm configurations and environmental data. With these pre-trained networks, we will then integrate RL to generate dynamic and safe arm-motion plans. The RL agent will leverage the IK and collision avoidance models to optimize arm trajectories, ensuring efficient and collision-free movements. This entire pipeline could be back propagated while promising to enhance the accuracy, safety, and flexibility of robotic arm operations in complex environments.
Keywords
Spot, Supervised learning, loco-manipulation
Labels
Master Thesis , ETH Zurich (ETHZ)
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-02-10 , Earliest start: 2025-02-10 , Latest end: 2026-03-01
Organization Robotic Systems Lab
Hosts Mirrazavi Sina
Topics Information, Computing and Communication Sciences
Model-Based Reinforcement Learning for Loco-manipulation
This project aims to develop a model-based reinforcement learning (RL) framework to enable quadruped robots to perform dynamic locomotion and manipulation simultaneously by leveraging advanced model-based RL algorithms such as DeamerV3, TDMPC2 and SAM-RL. We will develop control policies that can predict future states and rewards, enabling the robot to adapt its behavior on-the-fly. The primary focus will be on achieving stable and adaptive walking patterns while reaching and grasping objects. The outcome will provide insights into the integration of complex behaviors in robotic systems, with potential applications in service robotics and automated object handling.
Labels
Master Thesis , ETH Zurich (ETHZ)
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-02-10 , Earliest start: 2025-02-10 , Latest end: 2026-02-10
Organization Robotic Systems Lab
Hosts Mirrazavi Sina
Topics Information, Computing and Communication Sciences
Integrating OpenVLA for Vision-Language-Driven Loco-Manipulation robotics scenarios
This thesis proposes to integrate and adapt the OpenVLA (Open-Source Vision-Language-Action) model to control the Spot robotic arm for performing complex grasping and placing tasks. The study will focus on enabling the robot to recognize, grasp, and organize various toy-sized kitchen items based on human instructions. By leveraging OpenVLA's robust multimodal capabilities, this project aims to bridge the gap between human intent and robotic actions, enabling seamless task execution in unstructured environments. The research will explore the feasibility of fine-tuning OpenVLA for task-specific operations and evaluate its performance in real-world scenarios, providing valuable insights for advancing multimodal robotics.
Labels
Master Thesis , ETH Zurich (ETHZ)
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-02-10 , Earliest start: 2025-02-10 , Latest end: 2026-02-10
Organization Robotic Systems Lab
Hosts Mirrazavi Sina
Topics Information, Computing and Communication Sciences
Differentiable Simulation for Precise End-Effector Tracking
Unlock the potential of differentiable simulation on ALMA, a quadrupedal robot equipped with a robotic arm. Differentiable simulation enables precise gradient-based optimization, promising greater tracking accuracy and efficiency compared to standard reinforcement learning approaches. This project dives into advanced simulation and control techniques, paving the way for improvements in robotic trajectory tracking.
Keywords
Differentiable Simulation, Learning, ALMA
Labels
Semester Project , Bachelor Thesis , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-02-07 , Earliest start: 2025-01-27
Organization Robotic Systems Lab
Hosts Mittal Mayank , Schwarke Clemens , Klemm Victor
Topics Information, Computing and Communication Sciences
Modeling and Simulation for Earthwork in Digital Twin
In this work, we aim to build a digital twin of our autonomous hydraulic excavator, leveraging Mathworks technology for high-fidelity modeling. This will be used in the future to test and benchmark our learning-based controllers.
Keywords
Modeling, Hydraulics, Excavation, Industry
Labels
Semester Project , Master Thesis , ETH Zurich (ETHZ)
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-02-06 , Earliest start: 2025-03-03
Organization Robotic Systems Lab
Hosts Spinelli Filippo , Nan Fang
Topics Information, Computing and Communication Sciences , Engineering and Technology
Reinforcement Learning for Excavation Planning In Terra
We aim to develop a reinforcement learning-based global excavation planner that can plan for the long term and execute a wide range of excavation geometries. The system will be deployed on our legged excavator.
Keywords
Keywords: Reinforcement learning, task planning
Labels
Semester Project , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-02-03 , Earliest start: 2025-03-01 , Latest end: 2025-08-31
Organization Robotic Systems Lab
Hosts Terenzi Lorenzo
Topics Information, Computing and Communication Sciences
Model Based Reinforcement Learning
We want to train an excavator agent to learn in a variety of soil using a fast, GPU-accelerated soil particle simulator in Isaac Sim.
Keywords
particle simulation, omniverse, warp, reinforcement learning, model based reinforcement learning.
Labels
Semester Project , Master Thesis
Description
Work Packages
Requirements
More information
Open this project... call_made
Published since: 2025-02-03 , Earliest start: 2025-02-28 , Latest end: 2025-08-31
Organization Robotic Systems Lab
Hosts Egli Pascal Arturo , Terenzi Lorenzo
Topics Information, Computing and Communication Sciences , Engineering and Technology
Reinforcement Learning for Particle-Based Excavation in Isaac Sim
We want to train RL agents on our new particle simulator, accelerated on the GPU via warp in Isaac sim.
Keywords
particle simulation, omniverse, warp, reinforcement learning
Labels
Semester Project , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-02-03 , Earliest start: 2025-03-01 , Latest end: 2025-09-30
Organization Robotic Systems Lab
Hosts Egli Pascal Arturo , Mittal Mayank , Terenzi Lorenzo
Topics Information, Computing and Communication Sciences
Perceptive Reinforcement Learning for Exavation
In this project, our goal is to leverage precomputed embeddings(VAE in Isaacsim) from 3D earthworks scene reconstructions to train reinforcement learning agents. These embeddings, derived from incomplete point cloud data and reconstructed using an encoder-decoder neural network, will serve as latent representations. The main emphasis is on utilizing these representations to develop and train reinforcement learning policies for digging tasks.
Keywords
LIDAR, 3D reconstruction, Isaac gym, deep learning, perception, reinforcement learning
Labels
Semester Project , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-02-03 , Earliest start: 2025-01-01 , Latest end: 2025-08-31
Organization Robotic Systems Lab
Hosts Höller David , Terenzi Lorenzo
Topics Information, Computing and Communication Sciences
Reiforcement Learning of Pretrained Trasformer Models
We want to train RL agents on our new particle simulator, accelerated on the GPU via warp in Isaac sim.
Keywords
Keywords: particle simulation, omniverse, warp, reinforcement learning
Labels
Semester Project , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-02-03 , Earliest start: 2025-03-01 , Latest end: 2025-08-31
Organization Robotic Systems Lab
Hosts Terenzi Lorenzo
Topics Information, Computing and Communication Sciences
Multiagent Reinforcement Learning in Terra
We want to train multiple agents in the Terra environment, a fully end-to-end GPU-accelerated environment for RL training.
Keywords
multiagent reinforcement learning, jax, deep learning, planning
Labels
Semester Project , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-02-03 , Earliest start: 2025-03-01 , Latest end: 2025-07-31
Organization Robotic Systems Lab
Hosts Terenzi Lorenzo
Topics Information, Computing and Communication Sciences
Propose Your Own Robotics Challenge
This project invites you to step into the role of an innovator, encouraging you to identify challenges you are passionate about within the field of robotics. Rather than working on predefined problems, you will have the freedom to propose your own project ideas, address real-world issues, or explore cutting-edge topics. This project allows you to define your own research journey.
Keywords
Robotics, Research
Labels
Semester Project , Bachelor Thesis , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-01-28 , Earliest start: 2025-01-27
Organization Robotic Systems Lab
Hosts Schwarke Clemens , Bjelonic Filip , Klemm Victor
Topics Information, Computing and Communication Sciences
Data Driven Simulation for End-to-End Navigation
Investigate how neural rendering can become the backbone of comprehensive, next generation data-driven simulation
Keywords
Neural rendering, Simulation
Labels
Internship , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-01-24 , Earliest start: 2025-01-27
Organization Robotic Systems Lab
Hosts Kneip Laurent
Topics Information, Computing and Communication Sciences , Engineering and Technology
Evolving Minds: Neuroevolution for Legged Locomotion
This project explores the use of neuroevolution for optimizing control policies in legged robots, moving away from classical gradient-based methods like PPO. Neuroevolution directly optimizes network parameters and structures, potentially offering advantages in environments with sparse rewards, while requiring fewer hyperparameters to tune. By leveraging genetic algorithms and evolutionary strategies, the project aims to develop efficient controllers for complex locomotion tasks. With computational capabilities doubling approximately every two years as predicted by Moore's Law, neuroevolution offers a promising approach for scaling intelligent control systems.
Keywords
Evolutionary Algorithms, Reinforcement Learning, Quadrupeds, Legged Locomotion
Labels
Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-01-22 , Earliest start: 2025-01-27
Organization Robotic Systems Lab
Hosts Bjelonic Filip , Schwarke Clemens
Topics Information, Computing and Communication Sciences
Design of a Compliant Mechanism for Human-Robot Collaborative Transportation with Non-Holonomic Robots
Human-robot collaboration is an attractive option in many industries for transporting long and heavy items with a single operator. In this project, we aim to enable HRC transportation with a non-holonomic robotic base platform by designing a compliant manipulation mechanism, inspired by systems like the Omnid Mocobots.
Keywords
Human-robot collaboration Collaborative transportation Non-holonomic robot Mobile manipulation
Labels
Master Thesis , ETH Zurich (ETHZ)
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2025-01-16 , Earliest start: 2024-07-08
Applications limited to ETH Zurich
Organization Robotic Systems Lab
Hosts Kindle Julien , Bray Francesca
Topics Information, Computing and Communication Sciences , Engineering and Technology
How to Touch: Exploring Tactile Representations for Reinforcement Learning
Developing and benchmarking tactile representations for dexterous manipulation tasks using reinforcement learning.
Keywords
Reinforcement Learning, Dexterous Manipulation, Tactile Sensing
Labels
Semester Project , Bachelor Thesis , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2025-01-08 , Earliest start: 2024-12-15 , Latest end: 2025-06-01
Applications limited to ETH Zurich
Organization Robotic Systems Lab
Hosts Bhardwaj Arjun , Zurbrügg René
Topics Information, Computing and Communication Sciences
BEV meets Semantic traversability
Enable Birds-Eye-View perception on autonomous mobile robots for human-like navigation.
Keywords
Semantic Traversability, Birds-Eye-View, Localization, SLAM, Object Detection
Labels
Master Thesis , ETH Zurich (ETHZ)
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2024-12-18 , Earliest start: 2025-01-15 , Latest end: 2025-10-31
Organization Robotic Systems Lab
Hosts Gawel Abel
Topics Information, Computing and Communication Sciences , Engineering and Technology
Scene graphs for robot navigation and reasoning
Elevate semantic scene graphs to a new level and perform semantically-guided navigation and interaction with real robots at The AI Institute.
Keywords
Scene graphs, SLAM, Navigation, Spacial Reasoning, 3D reconstruction, Semantics
Labels
Master Thesis , ETH Zurich (ETHZ)
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2024-12-18 , Earliest start: 2025-01-15 , Latest end: 2025-10-31
Organization Robotic Systems Lab
Hosts Gawel Abel
Topics Information, Computing and Communication Sciences , Engineering and Technology
Digital Twin for Spot's Home
MOTIVATION ⇾ Creating a digital twin of the robot's environment is crucial for several reasons: 1. Simulate Different Robots: Test various robots in a virtual environment, saving time and resources. 2. Accurate Evaluation: Precisely assess robot interactions and performance. 3. Enhanced Flexibility: Easily modify scenarios to develop robust systems. 4. Cost Efficiency: Reduce costs by identifying issues in virtual simulations. 5. Scalability: Replicate multiple environments for comprehensive testing. PROPOSAL We propose to create a digital twin of our Semantic environment, designed in your preferred graphics Platform to be able to simulate Reinforcement Learning agents in the digital environment, to create a unified evaluation platform for robotic tasks.
Keywords
Digital Twin, Robotics
Labels
Semester Project , Master Thesis
Contact Details
More information
Open this project... call_made
Published since: 2024-12-17 , Earliest start: 2025-01-05
Applications limited to University of Zurich , ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization Computer Vision and Geometry Group
Hosts Blum Hermann , Portela Tifanny , Bauer Zuria, Dr. , Trisovic Jelena
Topics Information, Computing and Communication Sciences
KALLAX Benchmark: Evaluating Household Tasks
Motivation ⇾ There are three ways to evaluate robots for pick-and-place tasks at home: 1. Simulation setups: High reproducibility but hard to simulate real-world complexities and perception noise. 2. Competitions: Good for comparing overall systems but require significant effort and can't be done frequently. 3. Custom lab setups: Common but lead to overfitting and lack comparability between labs. Proposal ⇾ We propose using IKEA furniture to create standardized, randomized setups that researchers can easily replicate. E.g, a 4x4 KALLAX unit with varying door knobs and drawer positions, generating tasks like "move the cup from the upper right shelf into the black drawer." This prevents overfitting and allows for consistent evaluation across different labs.
Keywords
Benchmakr, Robotics, pick-and-place
Labels
Semester Project , Master Thesis
Contact Details
More information
Open this project... call_made
Published since: 2024-12-17 , Earliest start: 2025-01-06
Applications limited to University of Zurich , ETH Zurich , Swiss National Science Foundation , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization Computer Vision and Geometry Group
Hosts Blum Hermann , Bauer Zuria, Dr. , Zurbrügg René
Topics Information, Computing and Communication Sciences
Adapting to Injuries for Dexterous In-Hand Manipulation
Develop a reinforcement learning-based method for training adaptive policies for dexterous in-hand manipulation systems to deal with actuator failure on the fly.
Keywords
Dexterous Manipulation, Reinforcement Learning, Adaptive Learning
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2024-12-12 , Earliest start: 2024-12-16 , Latest end: 2025-06-01
Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization Robotic Systems Lab
Hosts Bhardwaj Arjun , Ma Yuntao
Topics Information, Computing and Communication Sciences
Extending Functional Scene Graphs to Include Articulated Object States
While traditional [1] and functional [2] scene graphs are capable of capturing the spatial relationships and functional interactions between objects and spaces, they encode each object as static, with fixed geometry. In this project, we aim to enable the estimation of the state of articulated objects and include it in the functional scene graph.
Keywords
scene understanding, scene graph, exploration
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2024-12-11
Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization Computer Vision and Geometry Group
Hosts Bauer Zuria, Dr. , Trisovic Jelena , Zurbrügg René
Topics Information, Computing and Communication Sciences , Engineering and Technology
Active Object Localization with Touch
Develop active exploration strategies for object identification and localization with tactile feedback.
Keywords
Dexterous Manipulation, Object Retrieval, Active Localization, Tactile Sensing
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2024-12-11 , Earliest start: 2024-12-15 , Latest end: 2025-06-01
Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization Robotic Systems Lab
Hosts Bhardwaj Arjun , Zurbrügg René
Topics Information, Computing and Communication Sciences
Learning-based object orientation prediction for handovers
Humans are exceptional at handovers. Besides timing and spatial precision, they also have a high-level understanding of how the other person wants to use the object that is handed over. This information is needed to hand over an object, such that it can be used directly for a specific task. While robots can reason about grasp affordances, the integration of this information with perception and control is missing.
Keywords
Robot-Human Handover, Human-Robot-Interaction, Mobile Manipulation, Robotics
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2024-12-10 , Earliest start: 2025-02-01 , Latest end: 2025-12-31
Organization Robotic Systems Lab
Hosts Scheidemann Carmen , Tulbure Andreea
Topics Information, Computing and Communication Sciences , Engineering and Technology
Evolve to Grasp: Learning Optimal Finger Configuration for a Dexterous Multifingered Hand
Use evolutionary algorithms with analytical force closure metrics to learn the optimal morphology of a dexterous hand.
Keywords
Evolutionary Algorithm, Machine Learning, Grasping, Robotics
Labels
Semester Project , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2024-12-09 , Earliest start: 2024-12-09 , Latest end: 2025-10-31
Applications limited to ETH Zurich
Organization Robotic Systems Lab
Hosts Church Joseph , Zurbrügg René
Topics Information, Computing and Communication Sciences
Differential Particle Simulation for Robotics
This project focuses on applying differential particle-based simulation to address challenges in simulating real-world robotic tasks involving interactions with fluids, granular materials, and soft objects. Leveraging the differentiability of simulations, the project aims to enhance simulation accuracy with limited real-world data and explore learning robotic control using first-order gradient information.
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2024-12-09 , Earliest start: 2025-01-01 , Latest end: 2025-12-31
Applications limited to ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization Robotic Systems Lab
Hosts Nan Fang , Ma Hao
Topics Engineering and Technology
Conformal Prediction for Distribution Shift Detection in Online Learning
This project investigates the use of conformal prediction for detecting distribution shifts in online learning scenarios, with a focus on robotics applications. Distribution shifts, arising from deviations in task distributions or changes in robot dynamics, pose significant challenges to online learning systems by impacting learning efficiency and model performance. The project aims to develop a robust detection algorithm to address these shifts, classifying task distribution shifts as outliers while dynamically retraining models for characteristic shifts.
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2024-12-09 , Earliest start: 2025-01-01 , Latest end: 2025-12-31
Organization Robotic Systems Lab
Hosts Ma Hao , Nan Fang
Topics Information, Computing and Communication Sciences , Engineering and Technology
Visual Language Models for Long-Term Planning
This project uses Visual Language Models (VLMs) for high-level planning and supervision in construction tasks, enabling task prioritization, dynamic adaptation, and multi-robot collaboration for excavation and site management. prioritization, dynamic adaptation, and multi-robot collaboration for excavation and site management
Keywords
Visual Language Models, Long-term planning, Robotics
Labels
Semester Project , Master Thesis
Description
Work Packages
Contact Details
More information
Open this project... call_made
Published since: 2024-12-06 , Earliest start: 2025-01-24 , Latest end: 2025-10-29
Organization Robotic Systems Lab
Hosts Terenzi Lorenzo
Topics Information, Computing and Communication Sciences
Diffusion-based Shared Autonomy System for Telemanipulation
Robots may not be able to complete tasks fully autonomously in unstructured or unseen environments, however direct teleoperation from human operators may also be challenging due to the difficulty of providing full situational awareness to the operator as well as degradation in communication leading to the loss of control authority. This motivates the use of shared autonomy for assisting the operator thereby enhancing the performance during the task. In this project, we aim to develop a shared autonomy framework for teleoperation of manipulator arms, to assist non-expert users or in the presence of degraded communication. Imitation learning, such as diffusion models, have emerged as a popular and scalable approach for learning manipulation tasks [1, 2]. Additionally, recent works have combined this with partial diffusion to enable shared autonomy [3]. However, the tasks were restricted to simple 2D domains. In this project, we wish to extend previous work in the lab using diffusion-based imitation learning, to enable shared autonomy for non-expert users to complete unseen tasks or in degraded communication environments.
Keywords
Imitation learning, Robotics, Manipulation, Teleoperation
Labels
Semester Project , ETH Zurich (ETHZ)
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2024-12-02 , Earliest start: 2024-11-01 , Latest end: 2025-11-01
Applications limited to ETH Zurich , University of Zurich
Organization Robotic Systems Lab
Hosts Elanjimattathil Aravind
Topics Information, Computing and Communication Sciences , Engineering and Technology
Lifelike Agility on ANYmal by Learning from Animals
The remarkable agility of animals, characterized by their rapid, fluid movements and precise interaction with their environment, serves as an inspiration for advancements in legged robotics. Recent progress in the field has underscored the potential of learning-based methods for robot control. These methods streamline the development process by optimizing control mechanisms directly from sensory inputs to actuator outputs, often employing deep reinforcement learning (RL) algorithms. By training in simulated environments, these algorithms can develop locomotion skills that are subsequently transferred to physical robots. Although this approach has led to significant achievements in achieving robust locomotion, mimicking the wide range of agile capabilities observed in animals remains a significant challenge. Traditionally, manually crafted controllers have succeeded in replicating complex behaviors, but their development is labor-intensive and demands a high level of expertise in each specific skill. Reinforcement learning offers a promising alternative by potentially reducing the manual labor involved in controller development. However, crafting learning objectives that lead to the desired behaviors in robots also requires considerable expertise, specific to each skill.
Keywords
learning from demonstrations, imitation learning, reinforcement learning
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Klemm Victor
Topics Information, Computing and Communication Sciences
Pushing the Limit of Quadruped Running Speed with Autonomous Curriculum Learning
The project aims to explore curriculum learning techniques to push the limits of quadruped running speed using reinforcement learning. By systematically designing and implementing curricula that guide the learning process, the project seeks to develop a quadruped controller capable of achieving the fastest possible forward locomotion. This involves not only optimizing the learning process but also ensuring the robustness and adaptability of the learned policies across various running conditions.
Keywords
curriculum learning, fast locomotion
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization Robotic Systems Lab
Hosts Li Chenhao , Bagatella Marco , Li Chenhao , Li Chenhao , Li Chenhao
Topics Engineering and Technology
Humanoid Locomotion Learning and Finetuning from Human Feedback
In the burgeoning field of deep reinforcement learning (RL), agents autonomously develop complex behaviors through a process of trial and error. Yet, the application of RL across various domains faces notable hurdles, particularly in devising appropriate reward functions. Traditional approaches often resort to sparse rewards for simplicity, though these prove inadequate for training efficient agents. Consequently, real-world applications may necessitate elaborate setups, such as employing accelerometers for door interaction detection, thermal imaging for action recognition, or motion capture systems for precise object tracking. Despite these advanced solutions, crafting an ideal reward function remains challenging due to the propensity of RL algorithms to exploit the reward system in unforeseen ways. Agents might fulfill objectives in unexpected manners, highlighting the complexity of encoding desired behaviors, like adherence to social norms, into a reward function. An alternative strategy, imitation learning, circumvents the intricacies of reward engineering by having the agent learn through the emulation of expert behavior. However, acquiring a sufficient number of high-quality demonstrations for this purpose is often impractically costly. Humans, in contrast, learn with remarkable autonomy, benefiting from intermittent guidance from educators who provide tailored feedback based on the learner's progress. This interactive learning model holds promise for artificial agents, offering a customized learning trajectory that mitigates reward exploitation without extensive reward function engineering. The challenge lies in ensuring the feedback process is both manageable for humans and rich enough to be effective. Despite its potential, the implementation of human-in-the-loop (HiL) RL remains limited in practice. Our research endeavors to significantly lessen the human labor involved in HiL learning, leveraging both unsupervised pre-training and preference-based learning to enhance agent development with minimal human intervention.
Keywords
reinforcement learning from human feedback, preference learning
Labels
Master Thesis
Description
Goal
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Chen Xin , Li Chenhao
Topics Information, Computing and Communication Sciences , Engineering and Technology
Online Safe Locomotion Learning in the Wild
Reinforcement learning (RL) can potentially solve complex problems in a purely data-driven manner. Still, the state-of-the-art in applying RL in robotics, relies heavily on high-fidelity simulators. While learning in simulation allows to circumvent sample complexity challenges that are common in model-free RL, even slight distribution shift ("sim-to-real gap") between simulation and the real system can cause these algorithms to easily fail. Recent advances in model-based reinforcement learning have led to superior sample efficiency, enabling online learning without a simulator. Nonetheless, learning online cannot cause any damage and should adhere to safety requirements (for obvious reasons). The proposed project aims to demonstrate how existing safe model-based RL methods can be used to solve the foregoing challenges.
Keywords
safe mode-base RL, online learning, legged robotics
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao
Topics Engineering and Technology
Autonomous Curriculum Learning for Increasingly Challenging Tasks
While the history of machine learning so far largely encompasses a series of problems posed by researchers and algorithms that learn their solutions, an important question is whether the problems themselves can be generated by the algorithm at the same time as they are being solved. Such a process would in effect build its own diverse and expanding curricula, and the solutions to problems at various stages would become stepping stones towards solving even more challenging problems later in the process. Consider the realm of legged locomotion: Training a robot via reinforcement learning to track a velocity command illustrates this concept. Initially, tracking a low velocity is simpler due to algorithm initialization and environmental setup. By manually crafting a curriculum, we can start with low-velocity targets and incrementally increase them as the robot demonstrates competence. This method works well when the difficulty correlates clearly with the target, as with higher velocities or more challenging terrains. However, challenges arise when the relationship between task difficulty and control parameters is unclear. For instance, if a parameter dictates various human dance styles for the robot to mimic, it's not obvious whether jazz is easier than hip-hop. In such scenarios, the difficulty distribution does not align with the control parameter. How, then, can we devise an effective curriculum? In the conventional RSL training setting for locomotion over challenging terrains, there is also a handcrafted learning schedule dictating increasingly hard terrain levels but unified with multiple different types. With a smart autonomous curriculum learning algorithm, are we able to overcome separate terrain types asynchronously and thus achieve overall better performance or higher data efficiency?
Keywords
curriculum learning, open-ended learning, self-evolution, progressive task solving
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization Robotic Systems Lab
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Bagatella Marco , Li Chenhao
Topics Engineering and Technology
Humanoid Locomotion Learning with Human Motion Priors
Humanoid robots, designed to replicate human structure and behavior, have made significant strides in kinematics, dynamics, and control systems. Research aims to develop robots capable of performing tasks in human-centric settings, from simple object manipulation to navigating complex terrains. Reinforcement learning (RL) has proven to be a powerful method for enabling robots to learn from their environment, enhancing their performance over time without explicit programming for every possible scenario. In the realm of humanoid robotics, RL is used to optimize control policies, adapt to new tasks, and improve the efficiency and safety of human-robot interactions. However, one of the primary challenges is the high dimensionality of the action space, where handcrafted reward functions fall short of generating natural, lifelike motions. Incorporating motion priors into the learning process of humanoid robots addresses these challenges effectively. Motion priors can significantly reduce the exploration space in RL, leading to faster convergence and reduced training time. They ensure that learned policies prioritize stability and safety, reducing the risk of unpredictable or hazardous actions. Additionally, motion priors guide the learning process towards more natural, human-like movements, improving the robot's ability to perform tasks intuitively and seamlessly in human environments. Therefore, motion priors are crucial for efficient, stable, and realistic humanoid locomotion learning, enabling robots to better navigate and interact with the world around them.
Keywords
motion priors, humanoid, reinforcement learning, representation learning
Labels
Master Thesis
Description
Contact Details
More information
Open this project... call_made
Published since: 2024-11-26
Organization ETH Competence Center - ETH AI Center
Hosts Li Chenhao , Li Chenhao , Li Chenhao , Li Chenhao
Topics Information, Computing and Communication Sciences
DigGPT: Large Language Models for Excavation Planning
Large language models (LLMs) have shown the first sparks of artificial general intelligence. We want to test if GPT 4.0 can solve excavation planning problems.
Keywords
GPT, Large Language Models, Robotics, Deep Learning, Reinforcement Learning
Labels
Semester Project , Master Thesis
Description
Work Packages
Requirements
Contact Details
More information
Open this project... call_made
Published since: 2024-11-21 , Earliest start: 2025-01-01 , Latest end: 2025-08-31
Organization Robotic Systems Lab
Hosts Terenzi Lorenzo
Topics Engineering and Technology
VR in Habitat 3.0
Motivation: Explore the newly improved Habitat 3.0 simulator with a special focus on the Virtual Reality Features. This project is meant to be an exploration task on the Habitat 3.0 simulator, exploring all the newly introduced features focusing specifically on the implementation of virtual reality tools for scene navigation. The idea is to extend these features to self created environments in Unreal Engine that build uppon Habitat
Labels
Semester Project
Contact Details
More information
Open this project... call_made
Published since: 2024-11-06 , Earliest start: 2024-01-08
Applications limited to University of Zurich , Swiss National Science Foundation , ETH Zurich , EPFL - Ecole Polytechnique Fédérale de Lausanne
Organization Computer Vision and Geometry Group
Hosts Blum Hermann , Bauer Zuria, Dr. , Sun Boyang , Zurbrügg René
Topics Information, Computing and Communication Sciences
Generalist Excavator Transformer
We want to develop a generalist digging agent that is able to do multiple tasks, such as digging and moving loose soil, and/or control multiple excavators. We plan to use decision transformers, trained on offline data, to accomplish these tasks.
Keywords
Offline reinforcement learning, transformers, autonomous excavation
Labels
Semester Project , Master Thesis
PLEASE LOG IN TO SEE DESCRIPTION
More information
Open this project... call_made
Published since: 2024-10-11 , Earliest start: 2025-03-01 , Latest end: 2025-08-31
Organization Robotic Systems Lab
Hosts Werner Lennart , Egli Pascal Arturo , Terenzi Lorenzo , Nan Fang , Zhang Weixuan
Topics Information, Computing and Communication Sciences
Student Theses in Industry
We have a large number of industry partners who search for excellent students to conduct their student theses at the company or at ETH but in close collaboration with them (joint supervision by industry and ETH).
Ammann Group (Switzerland)
The Ammann Group is a worldwide leader in the manufacture of mixing plants, machinery, and services in the construction industry, with core competence in road construction and landscaping as well as in the transport infrastructure.
We are collaborating with Ammann to automate construction equipment
Maxon (Switzerland)
Maxon develops and builds electric drive systems that are among the best in the world. Their drive systems can be found wherever extreme precision and the highest quality standards are indispensable – on Earth, and on Mars.
Shunk (Germany)
Legged Wheel Chair

This project aims at extending a dynamic simulation and locomotion controllers for a robotized wheelchair able to handle difficult terrains including stairs. This project will prepare the prototype phase coming next.
.
Note on plagiarism
We would like to suggest every student, irrespective of the type of project (Bachelor, Semester, Master, ...), to make himself/herself familiar with ETH rules regarding plagiarism