Algorithms exploiting light’s polarization boost resolution of commercial depth sensors 1,000-fold
MIT researchers have shown that by exploiting the polarization of light — the physical phenomenon behind polarized sunglasses and most 3-D movie systems — they can increase the resolution of conventional 3-D imaging devices as much as 1,000 times.
The technique could lead to high-quality 3-D cameras built into cellphones, and perhaps to the ability to snap a photo of an object and then use a 3-D printer to produce a replica.
Further out, the work could also abet the development of driverless cars.
“Today, they can miniaturize 3-D cameras to fit on cellphones,” says Achuta Kadambi, a PhD student in the MIT Media Lab and one of the system’s developers. “But they make compromises to the 3-D sensing, leading to very coarse recovery of geometry. That’s a natural application for polarization, because you can still use a low-quality sensor, and adding a polarizing filter gives you something that’s better than many machine-shop laser scanners.”
The researchers describe the new system, which they call Polarized 3D, in a paper they’re presenting at the International Conference on Computer Vision in December. Kadambi is the first author, and he’s joined by his thesis advisor, Ramesh Raskar, associate professor of media arts and sciences in the MIT Media Lab; Boxin Shi, who was a postdoc in Raskar’s group and is now a research fellow at the Rapid-Rich Object Search Lab; and Vage Taamazyan, a master’s student at the Skolkovo Institute of Science and Technology in Russia, which MIT helped found in 2011.
When polarized light gets the bounce
If an electromagnetic wave can be thought of as an undulating squiggle, polarization refers to the squiggle’s orientation. It could be undulating up and down, or side to side, or somewhere in-between.
Polarization also affects the way in which light bounces off of physical objects. If light strikes an object squarely, much of it will be absorbed, but whatever reflects back will have the same mix of polarizations that the incoming light did. At wider angles of reflection, however, light within a certain range of polarizations is more likely to be reflected.
This is why polarized sunglasses are good at cutting out glare: Light from the sun bouncing off asphalt or water at a low angle features an unusually heavy concentration of light with a particular polarization. So the polarization of reflected light carries information about the geometry of the objects it has struck.
This relationship has been known for centuries, but it’s been hard to do anything with it, because of a fundamental ambiguity about polarized light. Light with a particular polarization, reflecting off of a surface with a particular orientation and passing through a polarizing lens is indistinguishable from light with the opposite polarization, reflecting off of a surface with the opposite orientation.
This means that for any surface in a visual scene, measurements based on polarized light offer two equally plausible hypotheses about its orientation. Canvassing all the possible combinations of either of the two orientations of every surface, in order to identify the one that makes the most sense geometrically, is a prohibitively time-consuming computation.
Polarization plus depth sensing
To resolve this ambiguity, the Media Lab researchers use coarse depth estimates provided by some other method, such as the time a light signal takes to reflect off of an object and return to its source. Even with this added information, calculating surface orientation from measurements of polarized light is complicated, but it can be done in real-time by a graphics processing unit, the type of special-purpose graphics chip found in most video game consoles.
The researchers’ experimental setup consisted of a Microsoft Kinect — which gauges depth using reflection time — with an ordinary polarizing photographic lens placed in front of its camera. In each experiment, the researchers took three photos of an object, rotating the polarizing filter each time, and their algorithms compared the light intensities of the resulting images.
On its own, at a distance of several meters, the Kinect can resolve physical features as small as a centimeter or so across. But with the addition of the polarization information, the researchers’ system could resolve features in the range of tens of micrometers, or one-thousandth the size.
Read more: Making 3-D imaging 1,000 times better
The Latest on: 3-D imaging
via Google News
The Latest on: 3-D imaging
- Imaging uncovers secrets of medicine's mysterious ivory manikinson December 1, 2019 at 7:46 am
"Digitizing and 3D printing them will give visitors more access and opportunity to interact with the manikins and may also allow investigators to learn more about their history." Radiological Society ...
- GE Healthcare's New Imaging Tech and Intelligent Apps Unveiled at #RSNA19 will Deliver Hundreds of Millions in Cost Savings and Provider Productivityon December 1, 2019 at 4:10 am
By leveraging data from GE Healthcare's 3D CT HD images, Embo ASSIST with Virtual Injection is designed to provide ... used systemwide as well as in the command center to pinpoint opportunities to ...
- EOS imaging RECEIVES U.S. FDA 510(K) CLEARANCE FOR EOSedge™ ITS NEW GENERATION INNOVATIVE IMAGING SYSTEMon November 29, 2019 at 4:01 pm
View the full release here: https://www.businesswire.com/news/home/20191128005574/en/ EOS imaging (EOSI.PA) (Euronext, FR0011191766 - EOSI - Eligible PEA - PME), a leader in 2D/3D orthopedic medical ...
- Innovative particle standards and long-lived imaging for 2D and 3D dSTORMon November 29, 2019 at 2:20 am
Stable localisation precisions within the 10 nm-range were achieved for single virions and in cellulo 2D imaging of centrosomes, as well as their reliable reconstruction in 3D dSTORM. dSTORM is based ...
- 3D scanner could aid accuracy in surgery, have other medical useson November 27, 2019 at 12:23 pm
“We’ve been able to measure the shape of a surface in real-time, whether it’s a rigid or a deforming surface,” says Jonathan Kofman, director of the University’s Intelligent Human-Machine Systems and ...
- EOS imaging Announces the Upcoming Launch of EOSedge™, Its New Generation Imaging Systemon November 26, 2019 at 4:06 pm
View the full release here: https://www.businesswire.com/news/home/20191126005639/en/ EOS imaging (EOSI.PA) (Euronext, FR0011191766 - EOSI - Eligible PEA - PME), a leader in 2D/3D orthopedic medical ...
- 4D-CT angiography versus 3D-rotational angiography as the imaging modality for computational fluid dynamics of cerebral aneurysmson November 26, 2019 at 10:24 am
Patient-specific models can be segmented from various imaging modalities, which may influence the geometric output and thus hemodynamic results. This study aims to compare CFD results from aneurysm ...
- Biz Buzz: Adventist Health’s Women’s Imaging Center hosts ribbon cuttingon November 25, 2019 at 7:00 pm
Services at the Women’s Imaging Center will include 3D mammography, 2D mammography, biopsy, ultrasound, bone density screening, hereditary cancer syndrome testing, breast health education and support ...
- 3D Sensor Market ($ 5.49 Billion) 2018-2023 market share, Future Scope, Demands and Projected Market Growthon November 25, 2019 at 1:29 am
Numerous technological innovations like multi detector CT scanners are currently getting adopted in the 3D imaging technologies, among the radiologists and the physicians, for performing timely ...
- Bringing faster 3-D imaging for biomedical researcheson November 21, 2019 at 6:39 am
This new compressive sensing two-photon microscopy can be applied to 3-D imaging of the nerve distribution of living things or to monitoring activities from hundreds of neurons simultaneously. New ...
via Bing News