Talk MA-Talk: Video Objekt Segmentierung für Omnidirektionale Stereo Panoramen
Speaker(s): Fan Song
Talk MA-Talk: Functional Volumetric Rendering for Industrial Applications
Speaker(s): Jan-Christopher Schmidt
Talk BA-Talk: Bekämpfung von Motion Sickness in VR durch dynamische bipolare Galvanisch Vestibuläre Stimulation
Speaker(s): Max Hattenbach
Talk BA-Talk: Neuronales Rendering - Wahrnehmungsbasierte Auswertung der Tiefenwirkung in VR
Speaker(s): Yannic Rühl
Talk BA-Talk: Entwurf einer interaktiven Simulation zum Erlernen von Sternenkonstellationen in öffentlichen Planetarien
Speaker(s): Lars Richard
Talk Promotions-V-Vg: Guiding Visual Attention in Immersive Environments
Growing popularity of virtual reality (VR) technology, presenting content virtually all around a user, creates new challenges for digital content creators and presentation systems. In this dissertation we investigate how to support viewers to not miss important information when exploring unknown virtual environments. We examine different visual stimuli to guide viewers' attention towards predetermined target regions of surrounding environments. To best possibly maintain the original visual appearance of scenes, we aim for subtle visual modifications that operate as close as possible to viewers' perception threshold, while still providing effective guidance.
In a first approach, we identify issues of existing visual guidance stimuli to be effective in VR environments. For use in large field of view (FOV) head-mounted displays (HMDs), we derive techniques to handle perspective distortions, degradation of visual acuity in the peripheral visual field and target regions outside the initial FOV. An existing visual stimulus, originally conceived for desktop environments, is adapted accordingly and successfully evaluated in a perceptual study.
Subsequently the generalizability of these extending techniques is investigated, regarding different guidance methods and VR devices. For this, additional methods from related work are re-implemented and updated accordingly. Two comparable perceptual studies are conducted to evaluate their effectiveness within a consumer-grade HMD and in an immersive dome projection system covering almost the full human visual field. Regardless of the actual success rates, all of the tested methods show a measurable effect on participants' viewing behavior, indicating general applicability of our modification techniques for various guiding methods and VR systems.
Finally, a novel visual guidance method (SIBM) is created, specifically designed for immersive systems. It builds on contrary manipulations of the two stereoscopic frames in VR rendering systems, turning the inevitable overhead of double (per eye) rendering into an advantage that is not available in monocular systems. Moreover, exploiting our visual system's sensitivity for discrepancies in binocular visual input, it allows to noticeably reduce the required per-image contrast of the actual stimulus well below previous state-of-the-art.
Talk SEP-Abschluss: Massively distributed collaborative crowd input system for dome environments
Dome (Aufnahmestudio & Visualisierungslabor)
Präsentation der Ergebnisse des studentischen Softwareentwicklungspraktikums (SEP).
Talk BA-Talk: Implementing Dynamic Stimuli in VR Environments for Visual Perception Research
Dome (Aufnahmestudio & Visualisierungslabor)
Speaker(s): Mai Hellmann
Talk VASC Seminar: Reconstructing 3D Human Avatars from Monocular Images
Carnegie Mellon University, Pittsburgh, PA, USA
Speaker(s): Thiemo Alldieck
Statistical 3D human body models have helped us to better understand human shape and motion and already enabled exciting new applications. However, if we want to learn detailed, personalized, and clothed models of human shape, motion, and dynamics, we require new approaches that learn from ubiquitous data such as plain RGB-images and video. I will discuss recent advances in personalized body shape and clothing estimation from monocular video, from a few frames, and even from a single image. We developed effective methods to learn detailed avatars without the need for expensive scanning equipment. These methods are easy to use and enable personalized avatar creation for example for VR and AR applications. I will conclude my talk by outlining the next challenges in human shape reconstruction.
Talk MA-Talk: Occlusion Aware Iterative Optical Flow Refinement for High Resolution Images
Speaker(s): Alexander Manegold
In the field of optical flow estimation many different approaches exist. Mostof the newest published methods use some kind of Convolutional NeuralNetwork (CNN). These CNNs often have high graphics hardware requirements which scale with the size of the input images. High resolution imagesor panoramas can consequently often not be processed at full resolution. The PanoTiler offers an image tiling strategy that can be used to partially estimate the optical flow using arbitrary CNNs and then merge the individual flow tiles. Its advantage over simple tiling techniques is the utilization of multiple resolution levels which allows to find better matching tile pairs between source and target images. Although the original PanoTiler yields good optical flow results for most images, errors are sometimes introduced in higher resolution levels. To solve this issue, I extended the PanoTiler approach with regularization which incorporates the optical flow of all levels for the final result. Additionally, I introduce a new optical flow clustering method to the PanoTiler which mends a vulnerability that produces errorsin higher resolution levels.
To compare the results of optical flow estimation techniques, multiple benchmarks like Middlebury, KITTI 2015 or MPI Sintel were created. These benchmarks mostly contain ground truth optical flow for lower resolution images and not for high resolution images or even panoramas. Because it is challenging to get ground truth optical flow for real world images, I created a simple to follow protocol to create panoramas and their ground truth opticalflow from Unreal Engine 4. The optical flow is generated using a Python tool and it is based on stereo vision and depth render passes.
Talk Fluid Simulation - From Research to Market
Speaker(s): Matthias Teschner
Based on many years of research at the University of Freiburg, FIFTY2 Technology develops and markets PreonLab, a framework for Lagrangian fluid simulation with a particular focus on the automotive industry. This presentation discusses the respective evolution from a research project to a product. The first part introduces selected research results that contribute to the success of PreonLab. The second part discusses the technology transfer and aspects that affect the prosperity of a university spin-off.
Talk Physics in Graphics: Measuring from Images
PhoenixD Retreat, Schneverdingen
Speaker(s): Marcus Magnor
Computer graphics is all about devising highly efficient, hardware-optimized algorithms to numerically evaluate the equations governing our physical world. Areas of physics that reguarly fall prey to computer graphics range from classical and continuum mechanics to hydrodynamics, optics, and radiation transport. In my talk I will give a few examples and discuss how being able to efficiently solve the forward problem of simulating the physical behavior of real-world systems can be used to also tackle the inverse problem of estimating and measuring physical properties from images.
Talk What’s missing in Head-mounted VR Displays?
Speaker(s): Marcus Magnor
Thanks to competitively priced HMDs geared towards the consumer market, research in immersive displays and Virtual Reality has seen tremendous progress. Still, a number of challenges remain to make immersive VR experiences truly realistic. In my talk I will shwocase a number of research projects at TU Braunschweig that aim to enhance the immersive viewing experience by taking perceptual issues and real-world recordings into account.
Symposium Computer Vision Colloquium
- 09.10.2019 15:00
National and international experts present the latest research in Computer Vision
Talk Creating Real VR Experiences
Informationzentrum, Seminarraum G30
Speaker(s): Tobias Bertel
Talk MA-Talk: Image Acquisition Strategies for Time-variant Projection of Phase Shift Patterns using a CMOS Rolling Shutter Image Sensor
Informatikzentrum, Seminarraum G30
Speaker(s): Benedikt-M. Pfennig