Skip to content
View rzninvo's full-sized avatar
🏠
Working from home
🏠
Working from home

Block or report rzninvo

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
rzninvo/README.md

Hi there, I'm Roham 👋

Typing SVG

Email LinkedIn ETH Zürich UZH Profile views


🧑‍🔬 About Me

I'm an MSc student in Artificial Intelligence at ETH Zürich and the University of Zürich. My work lies at the intersection of computer vision, robotics, and language-grounded AI, aiming to give machines a richer understanding of the physical world, from indoor localization via natural language to whole-body imitation learning for humanoid robots.

  • 🔭  Currently:  Graduate Student Researcher at the ETH Computer Vision and Geometry Group, building HERMES SLAM (CVPR 2027 target) with Dr. Dániel Béla Baráth; also RL Engineer in the Humanoid Group of the ETH Robotics Club
  • 🌱  Exploring:  Embodied AI · Vision-Language Models · Sim-to-Real Transfer · 3D Scene Understanding
  • 🤝  Open to collaborate on:  Spatial reasoning, RL for robotics, multimodal perception
  • 💬  Ask me about:  SLAM, scene graphs, indoor localization, humanoid control
  • 📫  Reach me at:  rzendehdel@ethz.ch
  • 🌍  Based in:  Zürich, Switzerland 🇨🇭
  • 🗣️  Speaks:  English · Persian · German · Japanese

🔬 Current Research & Projects

🛰️ HERMES SLAM  CVPR 2027 (target)

ETH D-INFK · Computer Vision and Geometry Group  ·  Graduate Student Researcher (Mar 2026 – Present)

A state-of-the-art Dynamic SLAM system developed at CVG with Dr. Dániel Béla Baráth, advancing real-time localization and mapping in scenes with moving objects, where classical SLAM assumptions break down.

Visual SLAM 3D Vision Dynamic Scenes C++ · Python

🧭 LangLoc  ECCV 2026 (under review)

ETH CVG Lab × UZH AI/ML Group

The first pipeline for fine-grained indoor localization from natural language alone, without any camera input. A dual-branch GATv2 + CLIP architecture for scene retrieval (+8 pp Top-1 over SOTA), a visibility-based floor-grid scoring module that reaches ~1 m median error, and a Bayesian dialog module that drives median error down to 7 cm via targeted yes/no questions.

PyTorch CLIP GNNs Bayesian Inference

🌊 Dark Diversity for eDNA

ETH Ecosystem & Landscape Evolution Lab

A Transformer-based mask modeling framework (inspired by Pl@ntBERT) on phylogenetic embeddings to detect ecologically suitable but absent marine species, quantifying anthropogenic impact at scale. Also built an end-to-end FASTQ → DADA2 → GBIF pipeline in collaboration with marine ecologists.

Transformers Bioinformatics eDNA R · Python

🤖 Humanoid Whole-Body Imitation

ETH Robotics Club

Implementing TWIST in NVIDIA IsaacLab to train a Unitree G1 locomotion policy on AMASS / MoCap data with robust sim-to-real transfer. Fine-tuning NVIDIA GR00T N1.6 on teleoperation data and deploying SONIC for real-time whole-body control, alongside a low-cost VR-based teleoperation rig built with PICO headsets.

IsaacLab RL Sim2Real Unitree G1

🗺️ SpotMap  Boston Dynamics Spot

ETH CVG Lab

Onboard 3D perception for the Spot robot: dense RGB-D SLAM with TSDF volumetric fusion from a single monocular camera, plus a dynamic scene graph generator powered by OpenMask3D for open-vocabulary instance segmentation, letting the robot semantically understand and update its world during interaction.

SLAM TSDF OpenMask3D Scene Graphs

🥽 Conversational Indoor Navigation

Microsoft Spatial AI Lab

A voice-based navigation assistant for smart glasses that guides users with context-aware, landmark-based verbal cues instead of metric instructions. Combines image-based localization, semantic landmark extraction, and mesh-aligned grounding inside Habitat-Sim, evaluated on a 3D mesh of the ETH HG building.

Habitat-Sim VLMs Smart Glasses Spatial AI

🚁 Vision-Based Drone Pursuit

UZH Robotics & Perception Group

A PPO-based vision policy for an autonomous "camera drone", trained in Flightmare / Agilicious to track dynamic targets while keeping safe flight dynamics. Deployed on real hardware through a 60 Hz ROS control loop, with extensive domain randomization for robust sim-to-real transfer.

PPO Flightmare ROS Sim2Real


🛠️ Tech Stack

 Languages

Python C++ C C# R SQL

 Machine Learning & Deep Learning

PyTorch TensorFlow Hugging Face scikit-learn NumPy Pandas SciPy OpenCV

 Robotics & Simulation

ROS 2 NVIDIA Isaac Lab MuJoCo Unity Habitat-Sim

 Tools & Infrastructure

Docker CUDA TensorRT Linux Git Jupyter MySQL


📊 GitHub Stats

GitHub stats Top languages

GitHub streak

Contribution graph


"The best way to predict the future is to invent it."
Alan Kay

Pinned Loading

  1. CNSG CNSG Public

    ETH Zurich - 263-5905-00L Mixed Reality HS2025: Conversational Navigation for Smart Glasses project.

    Python 4 2

  2. Digital-Twin-of-a-Traffic-Scene-Using-RSU-and-AWSIM Digital-Twin-of-a-Traffic-Scene-Using-RSU-and-AWSIM Public

    B. Sc. Thesis concerning the semi-automatic creation of a digital twin of a traffic scene near Amirkabir University of Technology

    Python 30 3

  3. robotics_final_project robotics_final_project Public

    The final project of the Robotics course. A robot has to find it's way out of a maze with the VFH algorithm. This robot should also detect road lanes, detect traffic signs in a city setting and act…

    Python 6

  4. CI_NeuroEvolutionGame CI_NeuroEvolutionGame Public

    A NeuroEvolution Model getting high scores in the SnailJumper game

    Python 3

  5. MixedRealityETHZ/Conversational_Navigation_for_Smart_Glasses MixedRealityETHZ/Conversational_Navigation_for_Smart_Glasses Public

    Python 1 1

  6. mMirmohammadi/Datathon_2026 mMirmohammadi/Datathon_2026 Public

    Python 1