Vedant Dave

Hey! I'm a Ph.D. student at Cyber-Physical-Systems Lab at Montanuniversität Leoben in Austria and advised by Elmar Rueckert. My research focuses on unsupervised reinforcement learning, robust representation learning under noisy multimodal sensory conditions, and robotic manipulation. I am broadly interested in developing scalable, generalizable representations for embodied agents in real-world, uncertainty-prone environments. Currently, I am a Research Intern in the Machine Learning and Data Science Unit at the Okinawa Institute of Science and Technology (OIST), Japan.

I received my Master's degree in Automation and Robotics from Technische Universität Dortmund in 2021, focusing on Robotics. My thesis, entitled “Model-agnostic Reinforcement Learning Solution for Autonomous Programming of Robotic Motion”, was completed at Mercedes-Benz AG, where I implemented reinforcement learning for motion planning of manipulators in complex environments. Prior to this, I was fortunate to work as a research intern with Leonel Rozo at the Bosch Center for Artificial Intelligence, where I worked on Probabilistic Movement Primitives on Riemannian Manifolds.

 /   /   /   /   / 

profile photo

Publications and Preprints

(* indicates equal contribution)
Papers under review are not displayed here.
DisDP: Robust Imitation Learning via Disentangled Diffusion Policies
Pankhuri Vanjani, Paul Mattes, Xiaogang Jia, Vedant Dave, Rudolf Lioutikov
Reinforcement Learning Conference (RLC) 2025
RSS Workshop on Reliable Robotics 2025
paper

Disentangled Diffusion Policy (DisDP) is an imitation learning method that improves robustness to sensor noise and failures by integrating multi-view disentanglement into diffusion-based policies. By decomposing sensory inputs into shared (global) and private (sensor-specific) representations, DisDP preserves task-relevant features while remaining resilient to perturbations. Evaluations on RoboColosseum and Libero show that DisDP matches baseline performance under normal conditions and outperforms them under sensor variations.

EnvoDat: A Large-Scale Multisensory Dataset for Robotic Spatial Awareness and Semantic Reasoning in Heterogeneous Environments
Linus Nwankwo, Björn Ellensohn, Vedant Dave, Peter Hofer, Jan Forstner, Marlene Villneuve, Robert Galler, Elmar Rueckert
IEEE International Conference on Robotics and Automation (ICRA) 2025
project page / code / paper

EnvoDat is a large-scale, multi-modal dataset designed to enhance robot autonomy beyond urban environments. It includes 26 sequences from 13 scenes, 10 sensing modalities, 1.9 TB of data, and 89K annotations across 82 classes. It is collected in diverse settings like underground tunnels, natural fields, and modern indoor spaces under conditions such as high illumination, fog, rain, and zero visibility. It supports benchmarking for SLAM and supervised learning algorithms and aids in fine-tuning multimodal vision models.

Skill Disentanglement in Reproducing Kernel Hilbert Space
Vedant Dave, Elmar Rueckert
Annual AAAI Conference on Artificial Intelligence (AAAI) 2025
Intrinsically Motivated Open-ended Learning, NeurIPS 2024
paper

Hilbert Unsupervised Skill Discovery (HUSD) is an unsupervised reinforcement learning method that learns a diverse and distinct set of skills without relying on external rewards. By combining Integral Probability Metrics with \(f\)-divergence, HUSD enhances exploration and ensures skill separability. It outperforms existing algorithms on state-based benchmarks, providing a robust foundation for various downstream tasks.

M2CURL: Sample-Efficient Multimodal Reinforcement Learning via Self-Supervised Representation Learning for Robotic Manipulation
Fotios Lygerakis, Vedant Dave, Elmar Rueckert
IEEE International Conference on Ubiquitous Robots (UR) 2024 (Best Student Paper Award)
ProxyTouch Workshop, ICRA 2024
project page / paper

M2CURL builds on MViTac to improves RL by efficiently integrating visual and tactile representations. It accelerates learning in downstream manipulation tasks.

Multimodal Visual-Tactile Representation Learning through Self-Supervised Contrastive Pre-Training
Vedant Dave*, Fotios Lygerakis*, Elmar Rueckert
IEEE International Conference on Robotics and Automation (ICRA) 2024
project page / code / paper / video

MViTac is self-supervised approach that integrates vision and tactile modalities using contrastive learning. MViTac utilizes intra- and inter-modality relationships to learn a shared representation space. Experiments show that MViTac outperforms state-of-the-art methods in downstream tasks, which are performed by linear probing.

Can we infer the full-arm manipulation skills from tactile targets?
Vedant Dave, Elmar Rueckert
Workshop on Advances in Close Proximity Human-Robot Collaboration, Humanoids 2022
paper

This paper contains late-breaking results for TacProMPs, which learns and predicts complex arm movements based on tactile responses. Experiments were conducted on the real robot with a wide variety of objects.

Predicting full-arm grasping motions from anticipated tactile responses
Vedant Dave, Elmar Rueckert
IEEE-RAS International Conference on Humanoid Robots (Humanoids) 2022 (Oral Presentation)
paper / video

TacProMPs learns and predicts complex arm movements based on tactile responses, particularly for manipulating non-uniform objects, demonstrating adaptability diverse grasping scenarios.

Orientation Probabilistic Movement Primitives on Riemannian Manifolds
Leonel Rozo*, Vedant Dave*
Conference on Robot Learning (CoRL) 2022
project page / paper

This paper introduces a Riemannian formulation of ProMPs for encoding and retrieving quaternion trajectories, enabling full-pose robot motions in operational spaces. This method builds on Riemannian manifold theory and exploits multilinear geodesic regression for estimating the ProMPs parameters.

Industrial Collaborations

Stahl- und Walzwerk Marienhütte GmbH

Consultant
2022/03-2023/02

Predicting Yield Strength of different materials from production process. Designing Neural Networks and optimization. Found out error in measurement inaccuracy from data analysis.

Academic Collaborations

Robust Representations during Sensory Loss
Paper in Reinforcement Learning Conference (RLC) 2025

This approach integrates multi-view disentanglement into diffusion-based policies, enhancing robustness to sensor noise and failures. By decomposing sensory inputs into shared (global) and private (sensor-specific) representations, it preserves task-relevant features while remaining resilient to perturbations.

A Reinforcement Learning Approach for Decision-Making in Wells
Under Review!

This research evaluates hole conditioning operations in wellbore drilling, focusing on activities like circulation, reaming, and washing to ensure well integrity. An agent is trained with model-free RL coming from the online data from the real-world scenarios.

Physics-informed neural network for predicting Gibbs free energy
Under Review!

Coding and Supervision

A physics-informed neural network combined with the CALPHAD formalism predicts Gibbs energy in alloys by determining the Redlich-Kister parameter using novel descriptors. This method enhances CALPHAD parameterization, expediting materials development and phase stability determination with high accuracy and potential.

Green Facade
Paper Coming Soon!

Coding and Supervision

This study predicts the floor level from the distribution of TCEs (Li, Be, V, Ga, Ge, Nb, Sb, Te, Ta, Tl, Bi, and REYs) in Vienna's urban aerosol.

Teaching Experience

cs188 Tutorial, Introduction to Machine Learning Lab SS 2024
Teaching Assistant, Introduction to Python WS 2023
Teaching Assistant, Cyber-Physical Systems Lab WS 2022

Reviewing Services

cs188

2025: ICML, NeurIPS
2024: AISTATS, ICLR, NeurIPS, ECAI, CoRL, ICRA, IROS, BioRob
2023: ECAI, CoRL, IROS, RAL
2022: ECAI, CoRL, IROS, RAL, Humanoids

Media

osr commulity

Template.