Christen Millerdurai

I am a Researcher and a Ph.D. candidate at Augmented Vision group, Deutsches Forschungszentrum für Künstliche Intelligenz (DFKI), advised by Prof. Dr. Didier Stricker, Dr. Alain Pagani, and Dr. Vladislav Golyanik. My research topic is 3D human pose estimation (including hands) and 3D scene reconstruction from egocentric cameras.

Previously, I obtained my MS at Saarland University (SIC) advised by Prof. Dr. Christian Theobalt on the topic of "3D Human Motion Capture from Egocentric Event Streams" and my BE from St. Joseph's Institute of Technology advised by Prof. Dr. C. Gnana Kousalya on the topic of "Visually Impaired Navigation With Artificial Intelligence".

E-mail: Christen.Millerdurai@dfki.de
Google Scholar  /  Linkedin /  Github

Portrait of Christen Millerdurai

Deutsches Forschungszentrum für Künstliche Intelligenz (DFKI)

DFKI
(2024-)

4DQV

4DQV
(2022-)

SIC

SIC
(2021-2024)

SJIT

St.Joseph's Institute of Technology
(2014-2018)

News

[December 2025] Two papers got accepted in WACV 2026.
[June 2025] One paper got published in IJCV 2025.
[December 2024] One paper got accepted in WACV 2025.
[May 2024] EventEgo3D was invited to a poster session in EgoVis workshop at CVPR 2024.
[Feb 2024] One paper got accepted to CVPR 2024.
[Dec 2023] One paper got accepted to 3DV 2024 (Spotlight).

Research

My research focus lies at the intersection of Computer Vision, Computer Graphics and Robotics. In particular, I am interested in the following topics:

  • Human/Hand Pose Estimation
  • 3D Scene Reconstruction
  • Egocentric 3D Vision
  • Robotics 3D Vision


Publications

TalkingPose: Efficient Face and Gesture Animation with Feedback-guided Diffusion Model

Alireza Javanmardi, Pragati Jaiswal, Tewodros Amberbir Habtegebrial, Christen Millerdurai, Shaoxiang Wang, Alain Pagani, Didier Stricker
Winter Conference on Applications of Computer Vision (WACV), 2026
[Project page] [Paper] [Code]

Recent advancements in diffusion models have significantly improved the realism and generalizability of character-driven animation, enabling the synthesis of high-quality motion from just a single RGB image and a set of driving poses. Nevertheless, generating temporally coherent long-form content remains challenging. Existing approaches are constrained by computational and memory limitations, as they are typically trained on short video segments, thus performing effectively only over limited frame lengths and hindering their potential for extended coherent generation. To address these constraints, we propose TalkingPose, a novel diffusion-based framework specifically designed for producing long-form, temporally consistent human upper-body animations. TalkingPose leverages driving frames to precisely capture expressive facial and hand movements, transferring these seamlessly to a target actor through a stable diffusion backbone. To ensure continuous motion and enhance temporal coherence, we introduce a feedback-driven mechanism built upon image-based diffusion models.

Inpaint360GS: Efficient Object-Aware 3D Inpainting via Gaussian Splatting for 360° Scenes

Shaoxiang Wang, Shihong Zhang, Christen Millerdurai, Rüdiger Westermann, Didier Stricker, Alain Pagani
Winter Conference on Applications of Computer Vision (WACV), 2026
[Project page] [Paper] [Code]

Despite recent advances in single-object front-facing inpainting using NeRF and 3D Gaussian Splatting (3DGS), inpainting in complex 360◦ scenes remains largely underexplored. This is primarily due to three key challenges: (i) identifying target objects in the 3D field of 360° environments, (ii) dealing with severe occlusions in multi-object scenes, which makes it hard to define regions to inpaint, and (iii) maintaining consistent and high-quality appearance across views effectively. To tackle these challenges, we propose Inpaint360GS, a flexible 360◦ editing framework based on 3DGS that supports multi-object removal and high-fidelity inpainting in 3D space. By distilling 2D segmentation into 3D and leveraging virtual camera views for contextual guidance, our method enables accurate object-level editing and consistent scene completion. We further introduce a new dataset tailored for 360◦ inpainting, addressing the lack of ground truth object-free scenes. Experiments demonstrate that Inpaint360GS outperforms existing baselines and achieves state-of-the-art performance.

EventEgo3D++: 3D Human Motion Capture from a Head Mounted Event Camera

Christen Millerdurai, Hiroyasu Akada, Jian Wang, Diogo Luvizon, Alain Pagani, Didier Stricker, Christian Theobalt, Vladislav Golyanik
International Journal of Computer Vision (IJCV), 2025
[Project page] [Paper] [Code]

This paper is an extension of our previous work "EventEgo3D (CVPR 2024)". We tackle a new problem, i.e. 3D human motion capture from an egocentric monocular event camera with a fisheye lens. "EventEgo3D (CVPR 2024)" proposed the EE3D framework that is specifically tailored for learning with event streams in the LNES representation, enabling high 3D reconstruction accuracy. We upgrade this framework to a new EE3D++ framework for further performance improvement. We also introduce a new dataset, EE3D-W, in addition to EE3D-S and EE3D-R from "EventEgo3D (CVPR 2024)".

Uni-slam: Uncertainty-aware neural implicit slam for real-time dense indoor scene reconstruction

Shaoxiang Wang, Yaxu Xie, Chun-Peng Chang, Christen Millerdurai, Alain Pagani, Didier Stricker
Winter Conference on Applications of Computer Vision (WACV), 2025
[Project page] [Paper] [Code]

Neural implicit fields have recently emerged as a powerful representation method for multi-view surface reconstruction due to their simplicity and state-of-the-art performance. However, reconstructing thin structures of indoor scenes while ensuring real-time performance remains a challenge for dense visual SLAM systems. Previous methods do not consider varying quality of input RGB-D data and employ fixed-frequency mapping process to reconstruct the scene, which could result in the loss of valuable information in some frames. In this paper, we propose Uni-SLAM, a decoupled 3D spatial representation based on hash grids for indoor reconstruction. We introduce a novel defined predictive uncertainty to reweight the loss function, along with strategic local-to-global bundle adjustment. Experiments on synthetic and real-world datasets demonstrate that our system achieves state-of-the-art tracking and mapping accuracy while maintaining real-time performance.

EventEgo3D: 3D Human Motion Capture from Egocentric Event Streams

Christen Millerdurai, Hiroyasu Akada, Jian Wang, Diogo Luvizon, Christian Theobalt, Vladislav Golyanik
Computer Vision and Pattern Recognition (CVPR), 2024
[Project page] [Paper] [Code]

We tackle a new problem, i.e. 3D human motion capture from an egocentric monocular event camera with a fisheye lens. Event streams have high temporal resolution and could provide reliable cues for 3D human motion capture under high-speed human motions and rapidly changing illumination. We leverage these characteristics and propose the first approach for event-based 3D human pose estimation, EventEgo3D (EE3D). The proposed EE3D framework is specifically tailored for learning with event streams in the LNES representation, enabling high 3D reconstruction accuracy. We also provide two new datasets, EE3D-S and EE3D-R.

3D Pose Estimation of Two Interacting Hands from a Monocular Event Camera

Christen Millerdurai, Diogo Luvizon, Viktor Rudnev, André Jonas, Jiayi Wang, Christian Theobalt, Vladislav Golyanik
International Conference on 3D Vision (3DV), 2024, Spotlight
[Project page] [Paper] [Code]

3D hand tracking from monocular video is hard due to low-light and fast motion. RGB methods often fail in these cases due to motion blur and low-dynamic range. Event cameras avoid these issues but differ radically from images, so existing techniques don’t transfer. We present the first framework to track two fast, interacting hands from a single event camera, using a semi‑supervised feature‑wise attention module to resolve left‑right ambiguity and an intersection loss to prevent collisions. We release two datasets: Ev2Hands‑S (synthetic) and Ev2Hands‑R (real events with 3D ground truth). Our method improves 3D reconstruction accuracy and generalizes to real, low‑light conditions.

Experience & Education

Experience:

[Feb 2024 - Present] Researcher, DFKI (Kaiserslautern, Germany)


[Apr 2023 - Sep 2023] ML Intern, AUDI AG. (Ingolstadt, Germany)
[Sep 2022 - Mar 2023] Research Intern, Max Planck Institute for Informatics (Saarbrucken, Germany)
[Jun 2022 - Aug 2022] ML Intern, StarryAI.com. (Remote)
[Oct 2018 - Dec 2020] Engineer - II, Signzy. (Bangalore, India)
[Jun 2018 - Sep 2018] ML Intern, Signzy. (Bangalore, India)
[Jan 2018 - May 2018] Intern, Amazon.com, Inc. (Chennai, India)

Education:

[Feb 2024 - Present] Ph.D. student, DFKI (Kaiserslautern, Germany)
[Jan 2021 - Jan 2024] Master Student, Saarland University (Saarbrucken, Germany)
[Apr 2014 - May 2018] Bachelor student, St.Joseph's Institute of Technology (Chennai, India)

Professional Services & Activities

Thesis Students:

  • Darshankumar Asodariya (ongoing)
  • Sanjit Shashikant Chavan (Capturing and Detecting Micro Gestures Using Head-Mounted Devices, 2026)
  • Vippin Kumar Jeetmal (Multi-view Face and Gesture Animation with Dynamic Gaussians, 2025)

Teaching:

Reviewer Experience:

  • CVPR (2025-2026)
  • ECCV (2024-2026)
  • IEEE FG (2026)
  • NeurIPS (2025)
  • ICCV (2025)
  • Eurographics (2025)
  • IJCV (2025)
  • TPAMI (2025)
  • ACM TOG (2026)


© Christen Millerdurai / Design: jonbarron Icons: Freepik - Flaticon