Browsing by Author "Uchiyama, Hideaki"
Now showing 1 - 3 of 3
Results Per Page
Sort Options
Item 3D-Aware Image Relighting with Object Removal from Single Image(The Eurographics Association, 2022) Zhang, Yujia; Perusquia-Hernández, Monica; Isoyama, Naoya; Kawai, Norihiko; Uchiyama, Hideaki; Sakata, Nobuchika; Kiyokawa, Kiyoshi; Theophilus Teo; Ryota KondoWe propose a method to relight scenes in a single image while removing unwanted objects by the combination of 3D-aware inpainting and relighting for a new functionality in image editing. First, the proposed method estimates the depth image from an RGB image using single-view depth estimation. Next, the RGB and depth images are masked by the user by specifying unwanted objects. Then, the masked RGB and depth images are simultaneously inpainted by our proposed neural network. For relighiting, a 3D mesh model is first reconstructed from the inpainted depth image, and is then relit with a standard relighting pipeline. In this process, removing cast shadows and sky areas and albedo estimation are optionally performed to suppress the artifacts in outdoor scenes. Through these processes, various types of relighting can be achieved from a single photograph while excluding the colors and shapes of unwanted objects.Item A Data Collection Protocol, Tool and Analysis for the Mapping of Speech Volume to Avatar Facial Animation(The Eurographics Association, 2022) Miyawaki, Ryosuke; Perusquia-Hernandez, Monica; Isoyama, Naoya; Uchiyama, Hideaki; Kiyokawa, Kiyoshi; Hideaki Uchiyama; Jean-Marie NormandKnowing the relationship between speech-related facial movement and speech is important for avatar animation. Accurate facial displays are necessary to convey perceptual speech characteristics fully. Recently, an effort has been made to infer the relationship between facial movement and speech with data-driven methodologies using computer vision. To this aim, we propose to use blendshape-based facial movement tracking, because it can be easily translated to avatar movement. Furthermore, we present a protocol for audio-visual and behavioral data collection and a tool running on WEB that aids in collecting and synchronizing data. As a start, we provide a database of six Japanese participants reading emotion-related scripts at different volume levels. Using this methodology, we found a relationship between speech volume and facial movement around the nose, cheek, mouth, and head pitch. We hope that our protocols, WEB-based tool, and collected data will be useful for other scientists to derive models for avatar animation.Item Identifying Language-induced Mental Load from Eye Behaviors in Virtual Reality(The Eurographics Association, 2022) Schirm, Johannes; Perusquia-Hernández, Monica; Isoyama, Naoya; Uchiyama, Hideaki; Kiyokawa, Kiyoshi; Theophilus Teo; Ryota KondoWe compared content-independent eye tracking metrics under different levels of language-induced mental load in virtual reality (VR). We designed a virtual environment to balance consistent recording of eye data with user experience and freedom of action. We also took steps towards quantifying the phenomenon of not focusing exactly on surfaces by proposing ''focus offset'' as a VR-compatible eye metric. Responses to conditions with higher mental load included larger and more variable pupil sizes and less fixations. We also observed less voluntary gazing at distraction content, and a tendency of looking through surfaces.