objects are integrated into—and responsive to—the natural world. A virtual ball under your desk, for example, would be blocked from view unless you bent down to look at it. In theory, MR could become VR in a dark room.
• Augmented reality (AR) is a live direct or indirect view of a physical, real-world environment whose elements are augmented (or supplemented) by computer-generated sensory input such as sound, video, graphics or GPS data.
• Enhancing one’s current perception of reality with digital information and media, such as 3D models and videos • Overlaying in real-time the camera view of your smartphone, tablet, PC or connected glasses.
T Y E A R S • Focus on entreprise augmented reality apps • 2014: $247 million • 2019: $2.4 billion (prediction) • Consumer products launches expected for 2017 Source: https://www.juniperresearch.com/press/press-releases/enterprise-ar-app-revenues-reach-2-4bn-by-2019
S I O N : T Y P I C A L TA S K S • Recognition: Object recognition, Identification, Detection • Motion analysis: Egomotion, Tracking, Optical flow • Scene reconstruction: computing a S3 model of the scene • Image restoration: the removal of noises
S I O N : S Y S T E M M E T H O D S • Image acquisition • Pre-processing: re-sampling, noise reduction, constrast enhancement • Feature extraction • Detection / Segmentation • High-level processing: Image recognition / Image registration • Decision making
S I O N : I M A G E R E G I S T R AT I O N • First stage can use feature detection methods like corner detection, blob detection, edge detection or thresholding and/or other image processing methods. • The second stage restores a real world coordinate system from the data obtained in the first stage.
D P R O C E S S I N G • Converting the input image to grayscale • Performing binary threshold operations in order to generate a high contrast black and white image • Detecting contours in order to "bound" the marker • Identifying marker candidates, and then • Performing distortion correction in order to enable accurate marker decode
P R O C E S S I N G • Without fiducial markers, the camera position must be determined through “natural feature tracking” using feature-based detection, tracking, and matching. This approach is associated with the SLAM (simultaneous localization and mapping) techniques that have been developed in robotic research.
I T I O N • Augmented Reality Markup Language (ARML) is a data standard developed within the Open Geospatial Consortium (OGC), which consists of an XML grammar to describe the location and appearance of virtual objects in the scene, as well as ECMAScript bindings to allow dynamic access to properties of virtual objects.
O N C E P T S • Features represent the physical object that should be augmented. • VisualAssets describe the appearance of the virtual object in the augmented scene. • Anchors describe the spatial relation between the physical and the virtual object.
R An Anchor describes the location of the physical object in the real world. Four different Anchor types are defined in ARML: • Geometries • Trackables • RelativeTo • ScreenAnchor
O • Smartphones lack the understanding of the environment • Teach phone to see & understand the environment • Augment & improve our own ability to answer the question such as: • How many paints do I need to fill up this wall? • Will this couch fit in my living room? • How do I get from point A to point B?
E A R N I N G I N TA N G O ? Area Learning gives the device the ability to see and remember the key visual features of a physical space—the edges, corners, other unique features. • Drift correction (also called loop closures) • Area Description File (ADF) Align the virtual & the physical world.
P E R C E P T I O N I N TA N G O ? Common depth perception technologies: • Structured-light (require IR projector & IR sensor) • Time of flight (require IR projector & IR sensor) • Stereoscopy Tango's depth perception works best indoors at moderate distances (0.5 to 4 meters). Areas lit with light sources high in IR like sunlight or incandescent bulbs, or objects that do not reflect IR light, cannot be scanned well.
P E R C E P T I O N I N TA N G O ? Tango API provide a function to get depth data in the form of a point cloud. This format gives (x, y, z) coordinates for as many points in the scene as are possible to calculate. *{X, Y, Z, C}, C here means Confidence.
T H S E N S O R W O R K ? The IR projector projects a pattern of IR light which falls on objects around it like a sea of dots. We can't see the dots because the light is projected in the Infrared color range. The IR camera sees the dots and sends its video feed of this distorted dot pattern to the processor. Processor works out depth from the displacement of the dots: on near objects the pattern is spread out, on far objects the pattern is dense. Ref: https://jahya.net/blog/how-depth-sensor-works-in-5-minutes/
T E A R D O W N The depth-sensing array in Tango prototype includes: • an infrared projector • 4 MP rear-facing RGB/IR camera • 180º field of view fisheye rear-facing camera
T E A R D O W N IR projector: provides infrared light that other (non-RGB) cameras can use to get a sense of an area in 3D space. Quote Google: "The IR projector is from Mantis Vision, and designed specific to our specs for field of view and resolution. It is custom designed to work in partnership with the 4MP RGB-IR camera on the other side."
T E A R D O W N • The bright grid of dots shows that Tango works similarly to the original Microsoft Kinect, with a grid of dots to be captured by the IR sensors of the 4 MP camera, building a depth map.
E V I C E S • ARCore is designed to work on a wide variety of qualified Android phones running N and later. During the SDK preview, ARCore supports the following devices: • Google Pixel, Pixel XL, Pixel 2, Pixel 2 XL • Samsung Galaxy S8 (SM-G950U, SM-G950N, SM-G950F, SM-G950FD, SM-G950W, SM-G950U1)
S • Real life location & physical portal mapping • Collecting equipments / gears • Individual challenges (e.g. attack a portal & a stadium) • League / Fraction
F U T U R E We will have more and more ways to establish the communication between the virtual world and the physical world. When this day comes, technologies will be truly ubiquitous.
IMPORTANT: Send your slide (in PDF) to [email protected] with your team members’ names. They will be evaluated along with your presentation. • Any support is allowed (Slides, White board, paper drawing…) • Live demo is welcomed! • Be imaginative :) • Use lean canvas if you need help elaborating your ideas