Reconstruction of two dynamic scenes from monocular video; dynamic scenes show camera frustums indicating  recording trajectory of the input video

D-NPC: Dynamic Neural Point Clouds for Non-Rigid View Synthesis from Monocular Video

Oral Presentation Image-based renderingPoint-based modelsReconstructionRasterization

Eurographics (Computer Graphics Forum), 2025
Computer Graphics Lab
TU Braunschweig
Moritz Kappel,
Computer Graphics Lab
TU Braunschweig
Florian Hahlbohm,
Computer Graphics Lab
TU Braunschweig
Timon Scholz,
Computer Graphics Lab
TU Braunschweig
Susana Castillo,
Max Planck Institute for Informatics
Saarland Informatics Campus
Christian Theobalt,
Computer Graphics Lab
TU Braunschweig
Martin Eisemann,
Max Planck Institute for Informatics
Saarland Informatics Campus
Vladislav Golyanik,
Computer Graphics Lab
TU Braunschweig
Marcus Magnor

D-NPC enables fast, non-rigid view synthesis from monocular video by expanding extending Implicit Neural Point Clouds (INPC) to the temporal domain.

Reconstruction of two dynamic scenes from monocular video; dynamic scenes show camera frustums indicating  recording trajectory of the input video

Measuring Velocity Perception Regarding Stimulus Eccentricity

Oral Presentation VectionCybersicknessVirtual realityVelocity perceptionStimulus eccentricity

ACM Symposium on Applied Perception (SAP), 2024
Computer Graphics Lab
TU Braunschweig
Timon Scholz,
Computer Graphics Lab
TU Braunschweig
Colin Groth,
Computer Graphics Lab
TU Braunschweig
Susana Castillo,
Computer Graphics Lab
TU Braunschweig
Martin Eisemann,
Computer Graphics Lab
TU Braunschweig
Marcus Magnor

We investigate the influence of temporal frequency and eccentricity of a stimulus on the magnitude of perceived velocity in the periphery, building a model to predict the scaling factor for perceived velocity deviations relative to eccentricity.

Two pictures of a coral reef, showcasing the instability of traditional tone-mapping methods compared to our method highlighting significant improvements in color consistency.

Optimizing Temporal Stability in Underwater Video Tone Mapping

Poster PerceptionImage processing

Proc. Vision, Modeling and Visualization (VMV), 2023
Technische Universität
Braunschweig
Matthias Franz,
Technische Universität
Braunschweig
Bill Matthias Thang,
Technische Universität
Braunschweig
Pascal Sackhoff,
Computer Graphics Lab
TU Braunschweig
Timon Scholz,
Computer Graphics Lab
TU Braunschweig
Jannis Malte Möller,
Computer Graphics Lab
TU Braunschweig
Steve Grogorick,
Computer Graphics Lab
TU Braunschweig
Martin Eisemann

We present a method for temporal stabilization of depth-based underwater image tone mapping methods for application to monocular RGB video, improving color consistency across frames.

Person wearing a VR headset, reaching their hand out to touch a virtual object not visible  in the real world.

Instant Hand Redirection in Virtual Reality Through Electrical Muscle Stimulation-Triggered Eye Blinks

Hand redirectionEMSVirtual realityRedirectionVREye blinks

ACM Symposium on Virtual Reality Software and Technology (VRST), 2023
Computer Graphics Lab
TU Braunschweig
Colin Groth,
Computer Graphics Lab
TU Braunschweig
Timon Scholz,
Computer Graphics Lab
TU Braunschweig
Susana Castillo,
Computer Graphics Lab
TU Braunschweig
Jan-Philipp Tauscher,
Computer Graphics Lab
TU Braunschweig
Marcus Magnor

We propose a method for instant hand redirection in VR using electrical muscle stimulation (EMS) to trigger eye blinks, enabling seamless interaction between real and virtual objects.

© Website by Timon Scholz, 2025