Project Tango: The road to Google's VR ambitions
At the recently concluded Google I/O 2016, Project Tango was given a fair bit of attention at the developer sessions. While it's not new, just what is Project Tango about, and why would it matter more than VR in the short run? We cover these details as we try out the Yellowstone dev-kit running on the NVIDIA Tegra K1 processor.
By PY Ho -
Project Tango.
At the recently concluded Google I/O 2016, Daydream, the search giant’s VR solution, received much media attention. On the developer’s side of things however, Project Tango was given quite a fair bit of attention. There were almost as many Tango developer sessions as Daydream sessions. In fact, during developer sessions, Tango was frequently mentioned alongside with VR. Not to mention, Project Tango is one of the very few technologies where there is actual hardware with which developers can get started (as opposed to the bulk of the other announcements at I/O 2016). In fact, Lenovo just announced its first consumer-grade Tango device, the Phab 2 Pro.
So, what is Project Tango about and why it does it matter more than VR in the short run?
What is Tango?
In its basic form, Project Tango is a mobile technology platform whereby it utilizes specialized hardware and computer vision to accurately map, track, and measure a location without active use of GPS.
Devices?
Before 2016, there were only two dev-kit devices. The first device, codenamed ‘Peanut’, had a phone form factor, and was released in early 2014. However, what is still being sold online is the second device, codenamed 'Yellowstone'. It has a 7-inch table form factor, powered by NVIDIA Tegra K1 processor. Originally costing US$1,024, Google has since reduced the price to US$512.
The unit that the author owns is this ‘Yellowstone’ device.
Hardware at a glance
From the front, the Project Tango dev kit looks just like an ordinary 7-inch tablet. In fact, it looked like a more compact variant of the NVIDIA Shield Tablet.
The RGB-IR sensor on the left, the fisheye lens on the middle, and the depth sensor on the right.
The true power of the Tango device is on its back. Distinguishing it from other ordinary tablets are the presence of multiple sensors. The Yellowstone device has two cameras, of which one of them is a 4MP RGB-IR camera sensor. That might sound low resolution, but the camera can capture images, along with infrared information. IR data is important to assist with depth sensing. The other camera is a fisheye lens that primarily tracks motion. In addition to these cameras, there is an IR blaster for longer range depth sensing.
On the software side, the Tango tablet is still running on Android 4.4.2. Ancient indeed, but for developers of Tango apps, the focus is more on the Tango SDK than the Android OS.
Since the Tango tablet is more of a technology platform preview, we'll focus more on what Project Tango enables than the review of the hardware.
Fundamental features of Project Tango
There are three fundamental features of Project Tango:
Motion Tracking
Imagine you are blindfolded. You make movements. Sure, you know how hard you walked. But do you know how fast you have moved, or where you are now, relative to where you have started? Most modern mobile devices have accelerometers and gyroscopes, and hence, are able to track a user’s movement to a limited degree, such as hand movements and step counting. However, they do not perform well when tracking the user’s motion, relative to its starting location. As such, the only way to determine a user’s relative location and motion is by GPS.
Project Tango solves that by combining vision along with the motion sensors. In a nutshell, Tango estimates its current position by looking at the difference between two snapshots separated in a single timestep, combined with sensor information from the accelerometer and gyroscope. This allows Tango to record down the actual distance and speed travelled. This process is also known as visual-inertial odometry.
Area Learning
Area learning in action.
Despite the power of visual-inertial odometry, all devices cannot escape what is termed as sensor drift, whereby over time, the sensors will detect the device as drifting very slowly, even if it stationary. This effect is inherent in all movement sensors. As such, over a long period, motion tracking on an ordinary mobile device will deviate from its true motion by a big margin.
Tango has the capability to correct that by first learning the area, extracting various landmarks and features through a combination of visual and depth sensors, and stores it into memory. When the user navigates, Tango will offset the drift based on the landmarks and features previously extracted.
Such area learning can also be used to estimate a person’s position in an enclosed environment, without the need for hardware markers or GPS. Taking a human analogy, imagine you are trying to navigate streets, without using GPS. How do you recognize whether you have previously visited the place? You spot and remember various visual landmarks, such as a signboard or lamp post, such that you will know that you have passed by the same place while navigating around the streets.
Depth sensing
The inclusion of depth sensors in a Tango dev kit allows the device to perform 3D reconstruction of the surrounding environment. In addition, it also allows one to perform rough measurements of space, without needing a measuring tape. Here's an example:-
Here's a 3D reconstruction of a kitchen.
Given time, it is possible to smooth the 3D reconstruction.
Applications of Project Tango
Indoor localization
How does one find his or her location outdoors? By GPS usually. But what about indoors? With Tango’s ability to learn an area and recall, places of interest, such as museums, can deploy Tango to help visitors locate where exactly they are, or direct them to where they would like to go. In fact, Barcelona’s museum did exactly that.
True Augmented Reality (AR)
Augmented reality on existing mobile phone technology is currently limited to:-
- Relying on GPS to overlay information on landscape
- Overlay of 3D objects over a predefined marker
Due to lack of depth sensors, as well as limitations in computer vision techniques in determining surfaces and object edges, ordinary phones are not able to extract proper context of the environment, hence they are not able to properly size the overlaying 3D model for augmented reality.
Since Project Tango is capble of mapping out 3D space of an environment, it is able to accurately determine surfaces, and resize the overlaying 3D model to fit into the environment. Overlayed 3D models would then not look out of place.
Combined with motion tracking, the 3D model will stay in its place even when the user moves, thus allowing one to make use of true augmented reality to inspect the model closer.
This will give rise to applications in virtual home shopping, education, or simply just plain fun AR games.
What if I want to purchase and place this cube seat over here? I can now visualize this before making a decision.
AR + VR?
It is in practice possible. Just not Daydream compliant. Plus, the author does not have a 7-inch Google Cardboard
At I/O 2016, VR has hogged much spotlight in the keynote. VR in its current form however, has its limitations.
For one, you can't actually move within the VR world without an explicit form of input for movement. Put it simply, even if the VR user walks on his or her feet, he or she will still be stationery in the VR world - not counting HTC Vive as it uses room-scale sensors to bridge that limitation. But apart from that headset, the other VR headsets can't track relative motion yet (which is not the same has head tracking, which is supported). Also, there are limitations in implementing AR through VR headsets, because current VR headsets are not able to map 3D environments. This mixing of AR with VR is also termed as mixed reality - something that Microsoft's HoloLens is closest to achieving convincingly.
In the current state of VR, it’s just fully immersive virtual reality, with movements usually input by a controller - not really immersive, if one really thinks about it. You might say HTC Vive allows for movement, but at its current stage, it requires too much unblocked space and the headgear requires to be tethered to the PC, thus giving rise to backpack gaming machine concepts.
Project Tango overcomes these two problems of VR; as a Tango device, not only is it able to track motion, but it can also achieve true augmented reality. A user can actually move around in VR space purely by walking. He or she can even, for example, brandish a virtual sword, as the Tango is able to track features and overlay a 3D virtual sword over one's hands. That will then truly be immersive VR.
Closing Comments
Daydream may be Google’s conventional answer, or some say catch-up, to its competitors in the VR space, such as Microsoft’s HoloLens and Oculus Rift. However, Project Tango is probably the company's end-goal of how VR should actually work - VR without any physical controls and with mixed reality capability.
While this Project Tango dev kit is running on an old version of Android, Google has stated that Android N itself will support reading of sensors that are of six degree-of-freedom type (6DOF), as well as depth camera sensor type, both of which are fundamental ingredients for Project Tango. While there’s no word on whether the dev kit will ever receive Android N, we will see if other Tango devices, such as the Lenovo Phab 2 Pro, will have Android N baked in, and how it will fare against the dev kit. Stay tuned for a future update as we explore the Lenovo Phab 2 Pro.
Our articles may contain affiliate links. If you buy through these links, we may earn a small commission.