The Surface team today released a video highlighting the power of PixelSense.
"Microsoft's PixelSense, in the new Samsung SUR40 for Microsoft Surface, allows a display to recognize fingers, hands, and objects placed on the screen, enabling vision-based interaction without the use of cameras. The individual pixels in the display see what's touching the screen and that information is immediately processed and interpreted."
"Think of it like the connection between the eye and the brain. You need both, working together, to see. In this case, the eye is the sensor in the panel, it picks up the image and it feeds that to the brain which is our vision input processor that recognizes the image and does something with it. Taken in whole…this is PixelSense technology."
Here's how PixelSense actually works:
- A contact (finger/blob/tag/object) is placed on the display
- IR back light unit provides light (though the optical sheets, LCD and protection glass) that hits the contact.
- Light reflected back from the contact is seen by the integrated sensors.
- Sensors convert the light signal into an electrical signal/value.
- Values reported from all of the sensors are used to create a picture of what is on the display.
- The picture is analyzed using image processing techniques.
- The output is sent to the PC. It includes the corrected sensor image and various contact types (fingers/blobs/tags).
[Source:Microsoft Surface blog]