【开源方案共享】Google新开源AR:DepthLab

标题:DepthLab: Real-time 3D Interaction with Depth Maps for Mobile Augmented Reality

作者:Ruofei Du, Eric Turner, Maksym Dzitsiuk, Luca Prasso, Ivo Duarte, Jason Dourgarian, Joao Afonso, Jose Pascoal, Josh Gladstone, Nuno Cruces, Shahram Izadi, Adarsh Kowdle, Konstantine Tsotsos, David Kim† Google LLC

编译:particle

欢迎各位加入免费知识星球,获取PDF论文,欢迎转发朋友圈分享快乐。

本文介绍 Google最新开源的AR算法:DepthLab

代码:https://github.com/googlesamples/arcore-depth-lab

功能

  • 3D导向光标:渲染以屏幕为中心的三维光标。当三维光标沿物理曲面移动时,它应该根据曲面法线和距离更改其方向和比例。

  • 激光反射:通过触摸屏幕,沿相机主轴将虚拟激光从用户渲染到物理对象。激光到达表面时应该反射。hit和reflection算法对于移动AR开发人员应该是可重用的。

  • 物理测量:通过触摸手机屏幕上的像素来测量任意物理点的距离和高度(以米为单位)。

  • 化身移动:导航一个虚拟物体在物理环境中在两点之间自然移动。

  • 碰撞感知放置:测试虚拟对象的体积是否与观察到的环境曲面发生碰撞。

  • 虚拟阴影:渲染投射到物理曲面上的几何体感知阴影。阴影可以与任何具有虚拟对象的移动AR应用程序集成。

  • 环境纹理:使用其他材质(如熔岩、网格、草)重新纹理物理表面。这项技术也可以用来取代天花板的星图您的位置或生成一个地形与草,植物或岩石。

  • 物理仿真:模拟增强现实对象的物理现象,例如碰撞。

  • AR涂鸦:允许用户触摸屏幕,在实物上绘制/喷涂/绘制虚拟图纸。

等功能

英文摘要

Mobile devices with passive depth sensing capabilities are ubiquitous, and recently active depth sensors have become available on some tablets and AR/VR devices. Although real-time depth data is accessible, its rich value to mainstream AR applications has been sorely under-explored. Adoption of depth-based UX has been impeded by the complexity of performing even simple operations with raw depth data, such as detecting intersections or constructing meshes. In this paper, we introduce DepthLab, a software library that encapsulates a variety of depth-based UI/UX paradigms, including geometry-aware rendering (occlusion, shadows), surface interaction behaviors (physics-based collisions, avatar path planning), and visual effects (relighting, 3D-anchored focus and aperture effects). We break down the usage of depth into localized depth, surface depth, and dense depth, and describe our real-time algorithms for interaction and rendering tasks. We present the design process, system, and components of DepthLab to streamline and centralize the development of interactive depth features. We have open-sourced our software at https://github.com/googlesamples/arcore-depth-lab to external developers, conducted performance evaluation, and discussed how DepthLab can accelerate the workflow of mobile AR designers and developers. With DepthLab we aim to help mobile developers to effortlessly integrate depth into their AR experiences and amplify the expression of their creative vision.

资源

三维点云论文及相关应用分享

【点云论文速读】基于激光雷达的里程计及3D点云地图中的定位方法

3D目标检测:MV3D-Net

三维点云分割综述(上)

3D-MiniNet: 从点云中学习2D表示以实现快速有效的3D LIDAR语义分割(2020)

win下使用QT添加VTK插件实现点云可视化GUI

JSNet:3D点云的联合实例和语义分割

大场景三维点云的语义分割综述

PCL中outofcore模块---基于核外八叉树的大规模点云的显示

基于局部凹凸性进行目标分割

基于三维卷积神经网络的点云标记

点云的超体素(SuperVoxel)

基于超点图的大规模点云分割

更多文章可查看:点云学习历史文章大汇总

SLAM及AR相关分享

【开源方案共享】ORB-SLAM3开源啦!

【论文速读】AVP-SLAM:自动泊车系统中的语义SLAM

【点云论文速读】StructSLAM:结构化线特征SLAM

SLAM和AR综述

常用的3D深度相机

AR设备单目视觉惯导SLAM算法综述与评价

SLAM综述(4)激光与视觉融合SLAM

Kimera实时重建的语义SLAM系统

SLAM综述(3)-视觉与惯导,视觉与深度学习SLAM

易扩展的SLAM框架-OpenVSLAM

高翔:非结构化道路激光SLAM中的挑战

SLAM综述之Lidar SLAM

基于鱼眼相机的SLAM方法介绍

(0)

相关推荐