Julian Straub

Julian Straub

Julian Straub is a Research Scientist at Meta Reality Labs Research (RLR) working on Computer Vision and 3D Perception. Before joining RLR, Julian obtained his PhD on Nonparametric Directional Perception from MIT, where he was advised by John W. Fisher III and John Leonard within the CS and AI Laboratory (CSAIL). On his way to MIT, Julian graduated from the Technische Universität München (TUM) and the Georgia Institute of Technology with a M.Sc. He did his Diploma thesis in Eckehard Steinbach’s group with the NavVis founding team and in particular with Sebastian Hilsenbeck. At Georgia Tech Julian had the pleasure to work with Frank Daellart’s group.

Research

My current research interests are problems that involve 3D localization, recognition and description of all objects and surfaces from egocentric video streams in scalable and generalizable ways.

Here are some recent invited talks:

ShapeR: Robust Conditional 3D Shape Generation from Casual Captures
ShapeR: Robust Conditional 3D Shape Generation from Casual Captures

Yawar Siddiqui, Duncan Frost, Samir Aroudj, Armen Avetisyan, Henry Howard-Jenkins, Daniel DeTone, Pierre Moulon, Qirui Wu, Zhengqin Li, Julian Straub, Richard Newcombe, Jakob Engel

ShapeR generates high-fidelity 3D shapes from casually captured image sequences. It uses SLAM, 3D detection, and vision-language models to extract per-object conditioning and a rectified flow transformer to generate shapes, achieving 2.7x improvement in Chamfer distance over state of the art.

Aria Gen 2 Pilot Dataset
Aria Gen 2 Pilot Dataset

Chen Kong, James Fort, Aria Kang, Jonathan Wittmer, Simon Green, Tianwei Shen, Yipu Zhao, Cheng Peng, Gustavo Solaira, Andrew Berkovich, Nikhil Raina, Vijay Baiyya, Evgeniy Oleinik, Eric Huang, Fan Zhang, Julian Straub, Mark Schwesinger, Luis Pesqueira, Xiaqing Pan, Jakob Engel, Carl Ren, Mingfei Yan, Richard Newcombe

An egocentric multimodal dataset recorded with Aria Gen 2 glasses covering daily activities including cleaning, cooking, eating, playing, and outdoor walking. Includes raw sensor data and machine perception outputs.

DGS-LRM: Real-Time Deformable 3D Gaussian Reconstruction From Monocular Videos
DGS-LRM: Real-Time Deformable 3D Gaussian Reconstruction From Monocular Videos

Chieh Hubert Lin, Zhaoyang Lv, Songyin Wu, Zhen Xu, Thu Nguyen-Phuoc, Hung-Yu Tseng, Julian Straub, Numair Khan, Lei Xiao, Ming-Hsuan Yang, Yuheng Ren, Richard Newcombe, Zhao Dong, Zhengqin Li

A feed-forward method for real-time reconstruction of dynamic scenes from monocular video. DGS-LRM predicts deformable 3D Gaussians in a single forward pass, achieving results competitive with optimization-based methods while enabling long-range 3D object tracking.

Orienternet: Visual localization in 2d public maps with neural matching
Orienternet: Visual localization in 2d public maps with neural matching

Paul-Edouard Sarlin, Daniel DeTone, Tsun-Yi Yang, Armen Avetisyan, Julian Straub, Tomasz Malisiewicz, Samuel Rota Bulo, Richard Newcombe, Peter Kontschieder, Vasileios Balntas

We introduce the first deep neural network that can accurately localize an image using the same 2D semantic maps that humans use to orient themselves. OrienterNet leverages free and global maps from OpenStreetMap and is thus more accessible and more efficient than existing approaches.

Nonparametric Directional Perception
Nonparametric Directional Perception

Julian Straub

From an indoor scene to large-scale urban environments, a large fraction of man-made surfaces can be described by only a few planes with even fewer different normal directions. This sparsity is evident in the surface normal distributions, which exhibit a small number of concentrated clusters. In this work, I draw a rigorous connection between surface normal distributions and 3D structure, and explore this connection in light of different environmental assumptions to further 3D Perception.

Writeups

Community Service

Area Chair: CVPR 26

Reviewing:

* outstanding reviewer

Previous Mentees and Interns