Along with flying and invisibility, high on the list of every child’s aspirational superpowers is the ability to see through or around walls or other visual obstacles.
That capability is now a big step closer to reality as scientists from the University of Wisconsin–Madison and the Universidad de Zaragoza in Spain, drawing on the lessons of classical optics, have shown that it is possible to image complex hidden scenes using a projected “virtual camera” to see around barriers.
The technology is described in a report today (August 5, 2019) in the journal Nature. Once perfected, it could be used in a wide range of applications, from defense and disaster relief to manufacturing and medical imaging. The work has been funded largely by the military through the U.S. Defense Department’s Advanced Research Projects Agency (DARPA) and by NASA, which envisions the technology as a potential way to peer inside hidden caves on the moon and Mars.
Technologies to achieve what scientists call “non-line-of-sight imaging” have been in development for years, but technical challenges have limited them to fuzzy pictures of simple scenes. Challenges that could be overcome by the new approach include imaging far more complex hidden scenes, seeing around multiple corners and taking video.
“This non-line-of sight imaging has been around for a while,” says Andreas Velten, a professor of biostatistics and medical informatics in the UW School of Medicine and Public Health and the senior author of the new Nature study. “There have been a lot of different approaches to it.”
The basic idea of non-line of-sight imaging, Velten says, revolves around the use of indirect, reflected light, a light echo of sorts, to capture images of a hidden scene. Photons from thousands of pulses of laser light are reflected off a wall or another surface to an obscured scene and the reflected, diffused light bounces back to sensors connected to a camera. The recaptured light particles or photons are then used to digitally reconstruct the hidden scene in three dimensions.
“We send light pulses to a surface and see the light coming back, and from that we can see what’s in the hidden scene,” Velten explains.
Recent work by other research groups has focused on improving the quality of scene regeneration under controlled conditions using small scenes with single objects. The work presented in the new Nature report goes beyond simple scenes and addresses the primary limitations to existing non-line-of-sight imaging technology, including varying material qualities of the walls and surfaces of the hidden objects, large variations in brightness of different hidden objects, complex inter-reflection of light between objects in a hidden scene, and the massive amounts of noisy data used to reconstruct larger scenes.
Together, those challenges have stymied practical applications of emerging non-line-of-sight imaging systems.
Velten and his colleagues, including Diego Gutierrez of the Universidad de Zaragoza, turned the problem around, looking at it through a more conventional prism by applying the same math used to interpret images taken with conventional line-of-sight imaging systems. The new method surmounts the use of a single reconstruction algorithm and describes a new class of imaging algorithms that share unique advantages.
Conventional systems, notes Gutierrez, interpret diffracted light as waves, which can be shaped into images by applying well known mathematical transformations to the light waves propagating through the imaging system.
In the case of non-line-of-sight imaging, the challenge of imaging a hidden scene, says Velten, is resolved by reformulating the non-line-of-sight imaging problem as a wave diffraction problem and then using well-known mathematical transforms from other imaging systems to interpret the waves and reconstruct an image of a hidden scene. By doing this, the new method turns any diffuse wall into a virtual camera.
“What we did was express the problem using waves,” says Velten, who also holds faculty appointments in UW–Madison’s Department of Electrical and Computer Engineering and the Department of Biostatistics and Medical Informatics, and is affiliated with the Morgridge Institute for Research and the UW–Madison Laboratory for Optical and Computational Instrumentation. “The systems have the same underlying math, but we found that our reconstruction is surprisingly robust, even using really bad data. You can do it with fewer photons.”
Using the new approach, Velten’s team showed that hidden scenes can be imaged despite the challenges of scene complexity, differences in reflector materials, scattered ambient light and varying depths of field for the objects that make up a scene.
The ability to essentially project a camera from one surface to another suggests that the technology can be developed to a point where it is possible to see around multiple corners: “This should allow us to image around an arbitrary number of corners,” says Velten. “To do so, light has to undergo multiple reflections and the problem is how do you separate the light coming from different surfaces? This ‘virtual camera’ can do that. That’s the reason for the complex scene: there are multiple bounces going on and the complexity of the scene we image is greater than what’s been done before.”
According to Velten, the technique can be applied to create virtual projected versions of any imaging system, even video cameras that capture the propagation of light through the hidden scene. Velten’s team, in fact, used the technique to create a video of light transport in the hidden scene, enabling visualization of light bouncing up to four or five times, which, according to the Wisconsin scientist, can be the basis for cameras to see around more than one corner.
The technology could be further and more dramatically improved if arrays of sensors can be devised to capture the light reflected from a hidden scene. The experiments described in the new Nature paper depended on just a single detector.
In medicine, the technology holds promise for things like robotic surgery. Now, the surgeon’s field of view is restricted when doing sensitive procedures on the eye, for example, and the technique developed by Velten’s team could provide a more complete picture of what’s going on around a procedure.
In addition to helping resolve many of the technical challenges of non-line-of-sight imaging, the technology, Velten notes, can be made to be both inexpensive and compact, meaning real-world applications are just a matter of time.
The Latest on: Virtual camera
via Google News
The Latest on: Virtual camera
- The Best Camera Drones of 2019on August 15, 2019 at 10:01 am
Drone cameras vary widely, from simple video cameras designed to ... Some have VR capabilities, enabling you (if you have a virtual reality headset) to look around at how the world appears at 300 feet ...
- Chrome OS 76 released to stable with Android-style media controls, virtual desktop support, and moreon August 15, 2019 at 7:47 am
Virtual Desks on Chrome OS 76 ... Google didn't announce it, but some camera tweaks are also included in Chrome OS 76 (though we weren't able to see/confirm until the update landed on our devices).
- Most of this set is CGI, but it was all filmed in-camera in realtimeon August 15, 2019 at 7:03 am
The seamless melding of virtual sets with real cameras is on the rise, and the results are indistiguishable from reality. From a low budget filmmakers perspective, the idea that you could place your ...
- Snap launches Spectacles 3 with dual HD cameras and 3D - 9to5Macon August 14, 2019 at 5:20 pm
Snaps can also export to Camera Roll as circular, horizontal, square and virtual reality formats for saving, editing, sharing, and reliving, anywhere. Design Spectacles 3 feature a strong ...
- Snap unveils Spectacles 3 with 2 cameras to record in 3Don August 13, 2019 at 10:59 am
Snaps can also be exported to the camera rolls on users' smartphones as circular, horizontal, square and virtual reality formats for saving, editing and sharing outside of Snapchat. Spectacles 3 comes ...
- Snap Intros $380 Two-Camera Spectacles — A Fashion Accessory for Snapchat AR Contenton August 13, 2019 at 8:30 am
The Snaps also can be exported to the camera roll in circular, horizontal, square and virtual-reality formats. Spectacles 3 come in two colors: carbon, described as a “monochromatic black with a ...
- Snap debuts dual-camera Spectacles 3 glasses to create 3D AR contenton August 13, 2019 at 7:22 am
It says integrating dual cameras will enable Snaps to be taken in 3D and transformed with 3D effects, as well as viewed in virtual reality, and processed in “high definition” 2D with new ...
- New iPad Pros might come with a triple-lens cameraon August 10, 2019 at 8:21 am
Apple continues to make advances with AR and new, more powerful cameras would further allow its devices to better combine virtual objects with the real world. The new iPad Pros and 10.2-inch iPad ...
- NAU in Flagstaff will use cameras to implement "virtual parking" systemon August 9, 2019 at 9:47 am
FLAGSTAFF, Ariz. (AP) --Northern Arizona University plans to use cameras that read license plates to monitor parking rather than stickers affixed to windshields. The Arizona Daily Sun reported ...
- Northern Arizona University to implement ‘virtual parking’on August 8, 2019 at 10:32 pm
Northern Arizona University plans to use cameras that read license plates to monitor parking rather than stickers affixed to windshields, officials said. The "virtual parking" system will treat ...
via Bing News