<?xml version="1.0" encoding="utf-8"?>
<feed xmlns="http://www.w3.org/2005/Atom"><title>Titus Leistner</title><link href="https://titus-leistner.de/" rel="alternate"/><link href="https://titus-leistner.de/feeds/all.atom.xml" rel="self"/><id>https://titus-leistner.de/</id><updated>2025-08-10T00:00:00+02:00</updated><entry><title>A Mobile Scanning Solution to Reconstruct Strand-Based Hairstyles</title><link href="https://titus-leistner.de/a-mobile-scanning-solution-to-reconstruct-strand-based-hairstyles.html" rel="alternate"/><published>2025-08-10T00:00:00+02:00</published><updated>2025-08-10T00:00:00+02:00</updated><author><name>P. Grassal</name></author><id>tag:titus-leistner.de,2025-08-10:/a-mobile-scanning-solution-to-reconstruct-strand-based-hairstyles.html</id><summary type="html">&lt;p&gt;We propose a novel mobile scanning solution that allows end-users to reconstruct 3D hairstyles with actual per-strand curves from just a phone scan.&lt;/p&gt;</summary><content type="html">&lt;style&gt;
    .list-inline {
        padding-left: 0;
        list-style: none;
        margin-left: -5px;
        text-align: center;
    }

    .list-inline &gt; li {
        display: inline-block;
        padding-left: 5px;
        padding-right: 5px
    }
&lt;/style&gt;

&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        &lt;a href="https://philgras.github.io"&gt;P. Grassal&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://www.linkedin.com/in/luca-hormann-9431b5177/"&gt;L. Hormann&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://www.linkedin.com/in/nadia-hamlaoui-598461221/"&gt;N. Hamlaoui&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://titus-leistner.de"&gt;T. Leistner&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://www.linkedin.com/in/lynton-ardizzone/"&gt;L. Ardizzone&lt;/a&gt;
    &lt;/li&gt;
&lt;/ul&gt;
&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        Copresence
    &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://dl.acm.org/doi/10.1145/3721239.3734111"&gt;&lt;img alt="paper preview" src="https://titus-leistner.de/siggraph25.png"&gt;&lt;/a&gt;
This talk was presented by me at &lt;a href="https://s2025.siggraph.org"&gt;SIGGRAPH 2025&lt;/a&gt;.&lt;/p&gt;
&lt;h2&gt;Abstract&lt;/h2&gt;
&lt;p&gt;&lt;img alt="hair reconstruction pipeline" src="https://titus-leistner.de/hair_pipeline.png"&gt;
We propose a novel mobile scanning solution that allows end-users to reconstruct 3D hairstyles with actual per-strand curves from just a phone scan. Using a mixture of deep learning and optimization, we make hair scanning fast and accessible while delivering qualitative assets ready to be used in any 3D software.&lt;/p&gt;
&lt;h2&gt;Video&lt;/h2&gt;
&lt;video src="/videos/siggraph25.mp4" width="100%" controls="controls"&gt;&lt;/video&gt;

&lt;h2&gt;Citation&lt;/h2&gt;
&lt;div class="highlight"&gt;&lt;pre&gt;&lt;span&gt;&lt;/span&gt;&lt;code&gt;&lt;span class="nc"&gt;@inproceedings&lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="nl"&gt;grassal2025mobile&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;title&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{A Mobile Scanning Solution to Reconstruct Strand-Based Hairstyles}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;author&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Grassal, Philip-William and Hormann, Luca and Hamlaoui, Nadia and Leistner, Titus and Ardizzone, Lynton}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;booktitle&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Proceedings of the Special Interest Group on Computer Graphics and Interactive Techniques Conference Talks}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;pages&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{1--2}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;year&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{2025}&lt;/span&gt;
&lt;span class="p"&gt;}&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;</content><category term="Publications"/><category term="talk"/><category term="research"/><category term="computer vision"/><category term="avatars"/><category term="strand hair"/></entry><entry><title>PhD Thesis: Deep Learning-Based Depth Estimation from Light Fields</title><link href="https://titus-leistner.de/phd-thesis-deep-learning-based-depth-estimation-from-light-fields.html" rel="alternate"/><published>2024-07-23T00:00:00+02:00</published><updated>2024-07-23T00:00:00+02:00</updated><author><name>T. Leistner</name></author><id>tag:titus-leistner.de,2024-07-23:/phd-thesis-deep-learning-based-depth-estimation-from-light-fields.html</id><summary type="html">&lt;p&gt;Our paper on photorealistic human avatar generation.&lt;/p&gt;</summary><content type="html">&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        &lt;a href="https://titus-leistner.de"&gt;T. Leistner&lt;/a&gt;
    &lt;/li&gt;
&lt;/ul&gt;
&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        Heidelberg University
    &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://archiv.ub.uni-heidelberg.de/volltextserver/35163/1/inaugural_dissertation_leistner_titus.pdf"&gt;&lt;img alt="thesis preview" src="https://titus-leistner.de/thesis_short.png"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;Abstract&lt;/h2&gt;
&lt;p&gt;Light fields have emerged as a highly accurate method for depth estimation, known for its precision and robustness against occlusions.
After the decline of consumer-based light field cameras, new industrial and research applications have emerged with very different demands, including the usage of high-resolution wide-baseline camera arrays and the need for a reliable confidence measure.
This thesis responds to these evolving requirements with two main contributions:
First, the introduction of EPI-Shift, a deep learning-based framework for depth estimation from both, small- and wide-baseline light fields.
EPI-Shift combines discrete disparity classification with continuous disparity-offset regression and performs well on  wide-baseline light fields, even when trained solely on narrow-baseline data.
The second contribution focuses on multimodal posterior regression in depth estimation, useful for dealing with reflective and semi-transparent surfaces and for uncertainty quantification.
This thesis contributes three deep learning-based approaches for depth posterior regression: UPR, ESE, and DPP.
Each of these methods displays strengths and weaknesses for different applications, evaluated using a novel multimodal light field depth dataset.
Even with the extended applicability to wide-baseline light fields and the enhanced posterior regression capabilities, the performance of the presented methods stays on par with other state-of-the art approaches, marking a significant step towards practicality for today's applications.&lt;/p&gt;
&lt;h2&gt;Citation&lt;/h2&gt;
&lt;div class="highlight"&gt;&lt;pre&gt;&lt;span&gt;&lt;/span&gt;&lt;code&gt;&lt;span class="nc"&gt;@phdthesis&lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="nl"&gt;leistner2024deep&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;title&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Deep Learning-Based Depth Estimation from Light Fields}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;author&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Leistner, Titus}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;year&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{2024}&lt;/span&gt;
&lt;span class="p"&gt;}&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;</content><category term="Publications"/><category term="thesis"/><category term="research"/><category term="computer vision"/><category term="light fields"/></entry><entry><title>Towards Multimodal Depth Estimation from Light Fields</title><link href="https://titus-leistner.de/towards-multimodal-depth-estimation-from-light-fields.html" rel="alternate"/><published>2022-06-21T00:01:00+02:00</published><updated>2022-06-21T00:01:00+02:00</updated><author><name>T. Leistner</name></author><id>tag:titus-leistner.de,2022-06-21:/towards-multimodal-depth-estimation-from-light-fields.html</id><summary type="html">&lt;p&gt;My paper multimodal depth estimation.&lt;/p&gt;</summary><content type="html">&lt;style&gt;
    .list-inline {
        padding-left: 0;
        list-style: none;
        margin-left: -5px;
        text-align: center;
    }

    .list-inline &gt; li {
        display: inline-block;
        padding-left: 5px;
        padding-right: 5px
    }
&lt;/style&gt;

&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        &lt;a href="https://titus-leistner.de"&gt;T. Leistner&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://www.linkedin.com/in/radek-mackowiak-28856034"&gt;R. Mackowiak&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://scholar.google.com/citations?user=DsAf47EAAAAJ&amp;hl=en"&gt;L. Ardizzone&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://hci.iwr.uni-heidelberg.de/vislearn/people/ullrich-koethe/"&gt;U. Köthe&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://hci.iwr.uni-heidelberg.de/vislearn/people/carsten-rother/"&gt;C. Rother&lt;/a&gt;
    &lt;/li&gt;
&lt;/ul&gt;
&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        Heidelberg University
    &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/pdf/2203.16542.pdf"&gt;&lt;img alt="paper preview" src="https://titus-leistner.de/cvpr22_multimodal_short.png"&gt;&lt;/a&gt;
This paper was published at &lt;a href="https://cvpr2022.thecvf.com"&gt;CVPR 2022&lt;/a&gt;.
Download &lt;a href="https://titus-leistner.de/cvpr22_multimodal_slides.pdf"&gt;slides&lt;/a&gt; and &lt;a href="https://titus-leistner.de/cvpr22_multimodal_poster.pdf"&gt;poster&lt;/a&gt;.&lt;/p&gt;
&lt;h2&gt;Abstract&lt;/h2&gt;
&lt;p&gt;&lt;img alt="example of estimated posteriors" src="https://titus-leistner.de/multimodal_light_field_depth.png"&gt;
Light field applications, especially light field rendering and depth estimation, developed rapidly in recent years. While state-of-the-art light field rendering methods handle semi-transparent and reflective objects well, depth estimation methods either ignore these cases altogether or only deliver a weak performance. We argue that this is due current methods only considering a single "true" depth, even when multiple objects at different depths contributed to the color of a single pixel. Based on the simple idea of outputting a posterior depth distribution instead of only a single estimate, we develop and explore several different deep-learning-based approaches to the problem. Additionally, we contribute the first "multimodal light field depth dataset" that contains the depths of all objects which contribute to the color of a pixel. This allows us to supervise the multimodal depth prediction and also validate all methods by measuring the KL divergence of the predicted posteriors. With our thorough analysis and novel dataset, we aim to start a new line of depth estimation research that overcomes some of the long-standing limitations of this field.&lt;/p&gt;
&lt;h2&gt;Code&lt;/h2&gt;
&lt;p&gt;Our &lt;a href="https://github.com/titus-leistner/mmlf"&gt;implementation and dataset&lt;/a&gt; is hosted on GitHub.&lt;/p&gt;
&lt;h2&gt;Citation&lt;/h2&gt;
&lt;div class="highlight"&gt;&lt;pre&gt;&lt;span&gt;&lt;/span&gt;&lt;code&gt;&lt;span class="nc"&gt;@inproceedings&lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="nl"&gt;leistner2022towards&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;title&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Towards Multimodal Depth Estimation from Light Fields}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;author&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Leistner, Titus and Mackowiak, Radek and Ardizzone, Lynton and K{\&amp;quot;o}the, Ullrich and Rother, Carsten}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;booktitle&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;pages&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{12953--12961}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;year&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{2022}&lt;/span&gt;
&lt;span class="p"&gt;}&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;</content><category term="Publications"/><category term="paper"/><category term="research"/><category term="computer vision"/><category term="light fields"/></entry><entry><title>Neural Head Avatars from Monocular RGB Videos</title><link href="https://titus-leistner.de/neural-head-avatars-from-monocular-rgb-videos.html" rel="alternate"/><published>2022-06-21T00:00:00+02:00</published><updated>2022-06-21T00:00:00+02:00</updated><author><name>P. Grassal</name></author><id>tag:titus-leistner.de,2022-06-21:/neural-head-avatars-from-monocular-rgb-videos.html</id><summary type="html">&lt;p&gt;Our paper on photorealistic human avatar generation.&lt;/p&gt;</summary><content type="html">&lt;style&gt;
    .list-inline {
        padding-left: 0;
        list-style: none;
        margin-left: -5px;
        text-align: center;
    }

    .list-inline &gt; li {
        display: inline-block;
        padding-left: 5px;
        padding-right: 5px
    }
&lt;/style&gt;

&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        &lt;a href="https://philgras.github.io"&gt;P. Grassal&amp;sup1;*&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://www.linkedin.com/in/malte-prinzler/"&gt; M. Prinzler&amp;sup1;*&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://titus-leistner.de"&gt;T. Leistner&amp;sup1;&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://hci.iwr.uni-heidelberg.de/vislearn/people/carsten-rother/"&gt;C. Rother&amp;sup1;&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://www.niessnerlab.org/members/matthias_niessner/profile.html"&gt;M. Nießner&amp;sup2;&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://justusthies.github.io"&gt;J. Thies&amp;sup3;&lt;/a&gt;
    &lt;/li&gt;
&lt;/ul&gt;
&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        Heidelberg University&amp;sup1;
    &lt;/li&gt;
    &lt;li&gt;
        TU Munich&amp;sup2;
    &lt;/li&gt;
    &lt;li&gt;
        MPI for Intelligent Systems&amp;sup3;
    &lt;/li&gt;
&lt;/ul&gt;
&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        *equal contribution
    &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/pdf/2112.01554.pdf"&gt;&lt;img alt="paper preview" src="https://titus-leistner.de/cvpr22_neural_short.png"&gt;&lt;/a&gt;
This paper was published at &lt;a href="https://cvpr2022.thecvf.com"&gt;CVPR 2022&lt;/a&gt;.&lt;/p&gt;
&lt;h2&gt;Abstract&lt;/h2&gt;
&lt;p&gt;&lt;img alt="example of generated avatar" src="https://titus-leistner.de/neural_head_avatars.png"&gt;
We present Neural Head Avatars, a novel neural representation that explicitly models the surface geometry and appearance of an animatable human avatar that can be used for teleconferencing in AR/VR or other applications in the movie or games industry that rely on a digital human. Our representation can be learned from a monocular RGB portrait video that features a range of different expressions and views. Specifically, we propose a hybrid representation consisting of a morphable model for the coarse shape and expressions of the face, and two feed-forward networks, predicting vertex offsets of the underlying mesh as well as a view- and expression-dependent texture. We demonstrate that this representation is able to accurately extrapolate to unseen poses and view points, and generates natural expressions while providing sharp texture details. Compared to previous works on head avatars, our method provides a disentangled shape and appearance model of the complete human head (including hair) that is compatible with the standard graphics pipeline. Moreover, it quantitatively and qualitatively outperforms current state of the art in terms of reconstruction quality and novel-view synthesis.&lt;/p&gt;
&lt;h2&gt;Video&lt;/h2&gt;
&lt;video src="/videos/cvpr22_neural.mp4" width="100%" controls="controls"&gt;&lt;/video&gt;

&lt;h2&gt;Project Page and Code&lt;/h2&gt;
&lt;p&gt;Visit our &lt;a href="https://philgras.github.io/neural_head_avatars/neural_head_avatars.html"&gt;official project page&lt;/a&gt;.
Download or fork our &lt;a href="https://github.com/philgras/neural-head-avatars"&gt;implementation&lt;/a&gt; on GitHub.&lt;/p&gt;
&lt;h2&gt;Citation&lt;/h2&gt;
&lt;div class="highlight"&gt;&lt;pre&gt;&lt;span&gt;&lt;/span&gt;&lt;code&gt;&lt;span class="nc"&gt;@inproceedings&lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="nl"&gt;grassal2022neural&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;title&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Neural head avatars from monocular RGB videos}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;author&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Grassal, Philip-William and Prinzler, Malte and Leistner, Titus and Rother, Carsten and Nie{\ss}ner, Matthias and Thies, Justus}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;booktitle&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;pages&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{18653--18664}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;year&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{2022}&lt;/span&gt;
&lt;span class="p"&gt;}&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;</content><category term="Publications"/><category term="paper"/><category term="research"/><category term="computer vision"/><category term="avatars"/></entry><entry><title>Learning to Think Outside the Box: Wide-Baseline Light Field Depth Estimation with EPI-Shift</title><link href="https://titus-leistner.de/learning-to-think-outside-the-box-wide-baseline-light-field-depth-estimation-with-epi-shift.html" rel="alternate"/><published>2019-09-18T11:15:00+02:00</published><updated>2019-09-18T11:15:00+02:00</updated><author><name>T. Leistner</name></author><id>tag:titus-leistner.de,2019-09-18:/learning-to-think-outside-the-box-wide-baseline-light-field-depth-estimation-with-epi-shift.html</id><summary type="html">&lt;p&gt;My paper on light field depth estimation for large baseline camera arrays.&lt;/p&gt;</summary><content type="html">&lt;style&gt;
    .list-inline {
        padding-left: 0;
        list-style: none;
        margin-left: -5px;
        text-align: center;
    }

    .list-inline &gt; li {
        display: inline-block;
        padding-left: 5px;
        padding-right: 5px
    }
&lt;/style&gt;

&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        &lt;a href="https://titus-leistner.de"&gt;T. Leistner&amp;sup1;&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://www.linkedin.com/in/hendrik-schilling-a2019418a"&gt;H. Schilling&amp;sup1;&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://de.linkedin.com/in/radek-mackowiak-28856034"&gt;R. Mackowiak&amp;sup2;&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://tu-dresden.de/ing/informatik/smt/cgv/die-professur/inhaber-in?set_language=en"&gt;S. Gumhold&amp;sup3;&lt;/a&gt;
    &lt;/li&gt;
    &lt;li&gt;
        &lt;a href="https://hci.iwr.uni-heidelberg.de/vislearn/people/carsten-rother/"&gt;C. Rother&amp;sup1;&lt;/a&gt;
    &lt;/li&gt;
&lt;/ul&gt;
&lt;ul class="list-inline"&gt;
    &lt;li&gt;
        Heidelberg University&amp;sup1;
    &lt;/li&gt;
    &lt;li&gt;
        Robert Bosch GmbH&amp;sup2;
    &lt;/li&gt;
    &lt;li&gt;
        TU Dresden&amp;sup3;
    &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/pdf/1909.09059.pdf"&gt;&lt;img alt="paper preview" src="https://titus-leistner.de/3dv19_short.png"&gt;&lt;/a&gt;
This paper was published at &lt;a href="http://3dv19.gel.ulaval.ca"&gt;3DV 2019&lt;/a&gt; with an oral presentation.
Download &lt;a href="https://titus-leistner.de/3dv19_slides.pdf"&gt;slides&lt;/a&gt; and &lt;a href="https://titus-leistner.de/3dv19_poster.pdf"&gt;poster&lt;/a&gt;.&lt;/p&gt;
&lt;h2&gt;Abstract&lt;/h2&gt;
&lt;p&gt;&lt;img alt="basic idea of EPI-Shift" src="https://titus-leistner.de/epi_shift.png"&gt;
We propose a method for depth estimation from light field data, based on a fully convolutional neural network architecture.
Our goal is to design a pipeline which achieves highly accurate results for small- and wide-baseline light fields.
Since light field training data is scarce, all learning-based approaches use a small receptive field and operate on small disparity ranges.
In order to work with wide-baseline light fields, we introduce the idea of EPI-Shift:
To virtually shift the light field stack which enables to retain a small receptive field, independent of the disparity range.
In this way, our approach "learns to think outside the box of the receptive field".
Our network performs joint classification of integer disparities and regression of disparity-offsets.
A U-Net component provides excellent long-range smoothing.
EPI-Shift considerably outperforms the state-of-the-art learning-based approaches and is on par with hand-crafted methods.
We demonstrate this on a publicly available, synthetic, small-baseline benchmark and on large-baseline real-world recordings.&lt;/p&gt;
&lt;h2&gt;Results&lt;/h2&gt;
&lt;p&gt;&lt;img alt="experimental results" src="https://titus-leistner.de/epi_shift_results.png"&gt;&lt;/p&gt;
&lt;h2&gt;Code&lt;/h2&gt;
&lt;p&gt;Download or fork my &lt;a href="https://github.com/titus-leistner/epi-shift"&gt;implementation of EPI-Shift&lt;/a&gt; on GitHub.&lt;/p&gt;
&lt;h2&gt;Citation&lt;/h2&gt;
&lt;div class="highlight"&gt;&lt;pre&gt;&lt;span&gt;&lt;/span&gt;&lt;code&gt;&lt;span class="nc"&gt;@inproceedings&lt;/span&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="nl"&gt;leistner2019learning&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;title&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Learning to think outside the box: Wide-baseline light field depth estimation with EPI-shift}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="na"&gt;author&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{Leistner, Titus and Schilling, Hendrik and Mackowiak, Radek and Gumhold, Stefan and Rother, Carsten}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;booktitle&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{2019 International Conference on 3D Vision (3DV)}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;pages&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{249--257}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;year&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{2019}&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="w"&gt;    &lt;/span&gt;&lt;span class="na"&gt;organization&lt;/span&gt;&lt;span class="p"&gt;=&lt;/span&gt;&lt;span class="s"&gt;{IEEE}&lt;/span&gt;
&lt;span class="p"&gt;}&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;</content><category term="Publications"/><category term="paper"/><category term="research"/><category term="computer vision"/><category term="light fields"/></entry></feed>