Browsing by Author "Bloch, Isabelle"
Now showing 1 - 2 of 2
Results Per Page
Sort Options
Item Progressive and Efficient Multi-Resolution Representations for Brain Tractograms(The Eurographics Association, 2018) Mercier, Corentin; Gori, Pietro; Rohmer, Damien; Cani, Marie-Paule; Boubekeur, Tamy; Thiery, Jean-Marc; Bloch, Isabelle; Puig Puig, Anna and Schultz, Thomas and Vilanova, Anna and Hotz, Ingrid and Kozlikova, Barbora and Vázquez, Pere-PauCurrent tractography methods generate tractograms composed of millions of 3D polylines, called fibers, making visualization and interpretation extremely challenging, thus complexifying the use of this technique in a clinical environment. We propose to progressively simplify tractograms by grouping similar fibers into generalized cylinders. This produces a fine-grained multiresolution model that provides a progressive and real-time navigation through different levels of detail. This model preserves the overall structure of the tractogram and can be adapted to different measures of similarity. We also provide an efficient implementation of the method based on a Delaunay tetrahedralization. We illustrate our method using the Human Connectome Project dataset.Item Real-time Virtual-Try-On from a Single Example Image through Deep Inverse Graphics and Learned Differentiable Renderers(The Eurographics Association and John Wiley & Sons Ltd., 2022) Kips, Robin; Jiang, Ruowei; Ba, Sileye; Duke, Brendan; Perrot, Matthieu; Gori, Pietro; Bloch, Isabelle; Chaine, Raphaëlle; Kim, Min H.Augmented reality applications have rapidly spread across online retail platforms and social media, allowing consumers to virtually try-on a large variety of products, such as makeup, hair dying, or shoes. However, parametrizing a renderer to synthesize realistic images of a given product remains a challenging task that requires expert knowledge. While recent work has introduced neural rendering methods for virtual try-on from example images, current approaches are based on large generative models that cannot be used in real-time on mobile devices. This calls for a hybrid method that combines the advantages of computer graphics and neural rendering approaches. In this paper, we propose a novel framework based on deep learning to build a real-time inverse graphics encoder that learns to map a single example image into the parameter space of a given augmented reality rendering engine. Our method leverages self-supervised learning and does not require labeled training data, which makes it extendable to many virtual try-on applications. Furthermore, most augmented reality renderers are not differentiable in practice due to algorithmic choices or implementation constraints to reach real-time on portable devices. To relax the need for a graphics-based differentiable renderer in inverse graphics problems, we introduce a trainable imitator module. Our imitator is a generative network that learns to accurately reproduce the behavior of a given non-differentiable renderer. We propose a novel rendering sensitivity loss to train the imitator, which ensures that the network learns an accurate and continuous representation for each rendering parameter. Automatically learning a differentiable renderer, as proposed here, could be beneficial for various inverse graphics tasks. Our framework enables novel applications where consumers can virtually try-on a novel unknown product from an inspirational reference image on social media. It can also be used by computer graphics artists to automatically create realistic rendering from a reference product image.