Next we pry out this hunk of ... stuff, which turns out to be an earpiece speaker, mic, ambient light sensor (AMS TMD3702VC), and the Soli chip, for interpreting your gestures using the power of radar.
Google calls this implementation of its in-house Project Soli Motion Sense.
Although radar technology has been in use for a long time and seems simple enough on paper, we're at a loss as to how Google stuffed the entire system into a tiny featureless rectangle with no moving parts.
Motion Sense works by emitting precisely tuned waves of electromagnetic energy. When those waves bounce off of something (like your hand), some of them reflect back to the antenna.
The Soli chip then studies the reflected waves and analyzes their time delay, frequency shift, and other data to learn the characteristics of the object that reflected them—how big it is, how fast it's moving, in which direction, etc.
Soli then runs that data against its known gesture database to determine what action, if any, needs to be performed in the OS.
TL;DR: magic rectangle knows your every move.
Your contributions are licensed under the open source Creative Commons license.
Insert a new line below
Insert a new line above
Indent current line
Unindent current line
Delete current line
Navigate to next step.
Navigate to previous step.