Depth imaging system based on stereo vision and infrared radiation
09928420 · 2018-03-27
Assignee
Inventors
- Adam G. Kirk (Renton, WA, US)
- Oliver A. Whyte (Seattle, WA, US)
- Sing Bing Kang (Redmond, WA)
- Charles Lawrence Zitnick, III (Seattle, WA)
- Richard S. Szeliski (Bellevue, WA)
- Shahram Izadi (Cambridge, GB)
- Christoph Rhemann (Cambridge, GB)
- Andreas GEORGIOU (Cambridge, GB)
- Avronil Bhattacharjee (Redmond, WA)
Cpc classification
G06F3/0659
PHYSICS
G06F12/00
PHYSICS
B29C64/386
PERFORMING OPERATIONS; TRANSPORTING
B29C64/00
PERFORMING OPERATIONS; TRANSPORTING
G02B27/4233
PHYSICS
A63F13/213
HUMAN NECESSITIES
G06F11/3024
PHYSICS
G01B11/2545
PHYSICS
H04N13/239
ELECTRICITY
G01B11/2513
PHYSICS
G01B11/25
PHYSICS
H04N13/25
ELECTRICITY
H04N23/11
ELECTRICITY
G02B27/4205
PHYSICS
H04N13/254
ELECTRICITY
H04N2013/0081
ELECTRICITY
International classification
H04N13/00
ELECTRICITY
G06F12/00
PHYSICS
B29C64/00
PERFORMING OPERATIONS; TRANSPORTING
H04N17/00
ELECTRICITY
G01B11/25
PHYSICS
G06F9/30
PHYSICS
B29C64/386
PERFORMING OPERATIONS; TRANSPORTING
G02B27/42
PHYSICS
Abstract
The subject disclosure is directed towards a high resolution, high frame rate, robust stereo depth system. The system provides depth data in varying conditions based upon stereo matching of images, including actively illuminated IR images in some implementations. A clean IR or RGB image may be captured and used with any other captured images in some implementations. Clean IR images may be obtained by using a notch filter to filter out the active illumination pattern. IR stereo cameras, a projector, broad spectrum IR LEDs and one or more other cameras may be incorporated into a single device, which may also include image processing components to internally compute depth data in the device for subsequent output.
Claims
1. A system comprising: a projector configured to illuminate a scene with an unknown light pattern, the unknown light pattern emitted at a predetermined infrared (IR) wavelength; a first IR camera configured to capture a first image of the scene illuminated with the unknown light pattern; a second IR camera configured to capture a second image of the scene illuminated with the unknown light pattern; an optical notch filter configured to block the predetermined IR wavelength and pass other wavelengths of light; a third IR camera coupled to the optical notch filter, the third IR camera configured to capture a third image of the scene illuminated with the unknown light pattern, the third image comprising an IR image without the unknown light pattern; and an image processing system configured to compute depth data based upon the first, second, and third images captured by the first, second, and third IR cameras, and output the depth data from the device.
2. The system of claim 1 further comprising a visible light spectrum camera configured to capture a color image of the scene.
3. The system of claim 2, wherein the third IR camera and the visible light spectrum camera are combined into a single camera.
4. The system of claim 1 further comprising at least one IR light source configured to illuminate the scene with at least some IR light at a second IR wavelength that is not filtered out by the optical notch filter.
5. The system of claim 1, wherein at least one of the first, second, and third IR cameras is coupled to a narrow bandpass filter.
6. The system of claim 1, wherein at least two of the first, second, and third IR cameras are arranged as stereo IR cameras.
7. The system of claim 6, wherein the projector is positioned between the stereo IR cameras.
8. The system of claim 6 further comprising an additional camera configured to capture a fourth image of the scene, wherein the additional camera is positioned adjacent one of the first and second IR cameras.
9. The system of claim 6 further comprising an IR light source configured to illuminate the scene with at least some IR light at a second IR wavelength that is not filtered out by the optical notch filter, wherein the IR light source is positioned adjacent one of the first and second IR cameras, and the third IR camera is positioned adjacent the IR light source.
10. The system of claim 6 further comprising first and second IR light sources configured to illuminate the scene with at least some IR light at a second wavelength that is not filtered out by the optical notch filter, wherein the first IR light source is positioned adjacent one of the first and second IR cameras, and the third IR camera is positioned between the first IR light source and the second IR light source.
11. The system of claim 1 wherein first, second, and third IR cameras and projector are incorporated into a single device.
12. The system of claim 11 wherein the device includes an image processing system configured to compute the depth data within the device.
13. The system of claim 1 wherein the first, second, and third IR cameras and the projector are arranged linearly in a single device, and wherein the device comprises at least one IR light source configured to illuminate the scene with at least some IR light at a second wavelength that is not filtered out by the optical notch filter.
14. A system comprising: a plurality of infrared (IR) cameras configured to sense stereo images of a scene illuminated with an unknown IR light pattern, the unknown IR light pattern emitted at a predetermined IR wavelength; a plurality of narrow bandpass filters corresponding to the plurality of IR cameras, wherein a respective narrow bandpass filter of the plurality of narrow bandpass filters is coupled to a respective IR camera of the plurality of IR cameras, and wherein the plurality of narrow bandpass filters pass the predetermined IR wavelength and block other IR wavelengths of light; a notch-filtered camera configured with a notch filter to capture a clean image of the scene illuminated with the unknown IR light pattern, the clean image comprising an IR image without the unknown IR light pattern; and an image processor configured to compute depth data based upon the stereo images captured by the plurality of IR cameras and the clean image captured by the notch-filtered camera, and output the depth data from the device.
15. The system of claim 14 further comprising a projector configured to illuminate the scene with the unknown IR light pattern.
16. The system of claim 15, wherein the plurality of IR cameras, the notch-filtered camera, and the projector are arranged linearly in a device, and wherein the projector is positioned between two of the plurality of IR cameras.
17. The system of claim 14, wherein the notch-filtered camera configured with the notch filter is further configured to capture a visible light spectrum color image.
18. The system of claim 14 further comprising at least one IR light source that emits at least some IR light at a second wavelength that is not filtered out by the notch filter.
19. The system of claim 14, wherein the notch filter is configured to block the predetermined IR wavelength and pass visible and other IR wavelengths of light.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The present invention is illustrated by way of example and not limited in the accompanying figures in which like reference numerals indicate similar elements and in which:
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
DETAILED DESCRIPTION
(10) Various aspects of the technology described herein are generally directed towards various hardware configurations that are well-suited for a high resolution, high frame rate, robust stereo depth system. One implementation includes two infrared (IR) cameras, a patterned-light projector, and an optional RGB or RGB/IR camera (where IR includes near infrared, or NIR) with or without broad-spectrum IR lighting. These components may be built into a single device.
(11) Stereo images are captured from two cameras, using a projected light pattern to provide texture. There are general statistical properties of the light pattern, but in one embodiment, the exact pattern does not need to be known. As a result, the light projector need not be calibrated, which makes manufacturing cheaper. Further, the captured data do not degrade when another such device is projecting a pattern in the same room.
(12) In various implementations, having a light projector provides benefits, but in some implementations the light projector is not needed, whereby if the device is taken to an environment that degrades the SNR (signal-to-noise ratio) of the projector, such as outside of a building, the device still functions to an extent (e.g., provides depth data, but with possibly degraded quality).
(13) It should be understood that any of the examples herein are non-limiting. For example, while various camera and projector/emitter arrangements are exemplified herein, other arrangements may be used. Further, while RGB (red, green blue) color component data is described, data based upon other color schemes such as CMYK typically used in printing or 3D printing may be used. As such, the present invention is not limited to any particular embodiments, aspects, concepts, structures, functionalities or examples described herein. Rather, any of the embodiments, aspects, concepts, structures, functionalities or examples described herein are non-limiting, and the present invention may be used in various ways that provide benefits and advantages in active depth sensing, point clouds and image processing in general.
(14)
(15) In general, the two IR cameras 101 and 102 may be used for stereoscopic depth solving. The patterned-light projector 105 emits a pattern with statistical properties that are desirable for stereo matching.
(16) In some implementations such as the one represented in
(17) An example device layout may be that represented in
(18) Continuing with this example, depth may be computed from the IR camera 102, closest to the RGB/IR camera 103. The RGB/IR camera 103 may be positioned as close to the IR camera 102 as the LED 106 will allow, and the other LED 107 may be as close as possible to the RGB/IR camera 103. The RGB/IR camera 103 has a notch filter 104 that removes the projected pattern from the image captured by the RGB/IR camera 103, e.g., to provide clean images.
(19) Note that one otherwise generally identical arrangement may have only an RGB camera instead of the notch-filtered RGB/IR camera. Yet another otherwise generally identical arrangement may have only a notch-filtered IR camera, that is, without RGB.
(20) It is understood that the order/arrangement of these components is only one example arrangement, and that other arrangements may be used, including some of those exemplified hereinafter. Thus,
(21) In the example of
(22) In
(23) The LEDs 106 and 107 generally illuminate the scene with IR light. This allows capturing a clean IR image in conditions where ambient IR light is otherwise low. While optional, if an RGB/IR camera (with lighting if needed) is present, an IR image without the light pattern (which is optically filtered) is captured. This may be useful for face and object recognition in environments with no visible light. At the same time, the IR stereo cameras 221 and 222 (
(24) In
(25) The frames of images 114 and 115 captured by the cameras 101-103 are provided to an image processing system or subsystem 118. The image processing system or subsystem 118 includes a processor 120 and a memory 122, containing one or more image processing algorithms implemented in hardware/firmware logic and/or computer instructions, including a pixel matching/depth processing algorithm 124, which in general, outputs depth data 126, e.g., a depth map per frame.
(26) In some implementations, the image processing system 118 and image capturing system or subsystem 108, or parts thereof, may be combined into a single device, represented by the dashed box 130. For example a home entertainment device may include all of the components shown in
(27) Also shown in
(28) With the above arrangement, the IR cameras are able to capture stereo images used for depth determination. For example, as described in U.S. published patent application no. 20130100256, hereby incorporated by reference, different dots or other projected elements have different features when captured, including intensity (brightness), depending on the distance from the projector to the reflective surfaces and/or the distance from the camera to the reflective surfaces. As is also known, the projected texture pattern (projected dots are one example used herein) in different images taken at the same time (e.g., with genlocked stereo cameras) may be correlated with one another, such as by matching small (e.g., RGB) patches between RGB images of the same scene captured at the same instant. Thus, with captured images from IR cameras calibrated (e.g., for triangulation), known algorithms can determine depth maps using disparities of certain features between matched dots or local intensity distribution. This is one way in which a depth map may be obtained via stereo image processing.
(29) In the above arrangement, the projector need not be calibrated, only the cameras. This provides benefits over systems needing known projection pattern, as only needing to calibrate cameras is much easier, and the device can maintain calibration automatically. Notwithstanding, one or alternatives also may use known projection patterns.
(30)
(31)
(32)
(33)
(34) A light source may be calibrated with each camera, or some subset thereof. Each camera may thus provide a light sourcecamera depth map, as well as depth maps computed between them. For example, with two IR cameras IR1 and IR2 calibrated with a projector P, three depth maps IR1 and P, IR2 and P and IR1 and IR2 are available, which may be used to reduce errors, for example.
(35) In yet another possible implementation, a single LED is shown in a device 440 of
(36) Turning to another aspect, the IR cameras are generally positioned a relatively close distance apart from one another, so as to determine depths with good accuracy in a typical room, e.g., from 0.5 m to 5.0 m, for example. However, other scenarios may call for detecting depths at further distances, whereby a wider separation of the cameras provides for better triangulation. In
(37)
(38) In another aspect,
(39) The example in
(40) The optical path sharing mechanism 550 may be based upon reflection, e.g., via mirrors. As another alternative, an optical path may be the same for an actively illuminated image and a non-actively illuminated image by having one camera 660 (
(41) The system can be adapted to various scenarios by providing a device with the components that are appropriate for a given scenario. Further, the system itself can adapt. For example, in some situations LED IR illumination, and/or active pattern illumination may be turned on or off as needed. Depth sensing by processing stereo images may still be performed regardless of whether active projection is turned off or on but is overwhelmed by ambient light. One or more configurations allow for both accurate depth at both close/short range and far long-range depth by having multiple image pairs at different perspectives to choose from in stereo matching.
(42) Further, unlike calibrated projector and light sensors, some of the implementations described herein as well as others not explicitly described may benefit from multiple devices, or at least multiple projectors in the same room, as more texture is added by more projectors. Even if the number of projectors is such that textures begin to blur, however, depth sensing may occur via processing the stereo images and/or the RGB/IR clean image. In other implementations, different wavelength projectors along with corresponding filtering may be used to capture only as much active illumination as desired.
(43) In another implementation generally represented in
(44) Example Operating Environment
(45)
(46) The invention is operational with numerous other general purpose or special purpose computing system environments or configurations. Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the invention include, but are not limited to: personal computers, server computers, hand-held or laptop devices, tablet devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
(47) The invention may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, and so forth, which perform particular tasks or implement particular abstract data types. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in local and/or remote computer storage media including memory storage devices.
(48) With reference to
(49) The computer 810 typically includes a variety of computer-readable media. Computer-readable media can be any available media that can be accessed by the computer 810 and includes both volatile and nonvolatile media, and removable and non-removable media. By way of example, and not limitation, computer-readable media may comprise computer storage media and communication media. Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by the computer 810. Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above may also be included within the scope of computer-readable media.
(50) The system memory 830 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 831 and random access memory (RAM) 832. A basic input/output system 833 (BIOS), containing the basic routines that help to transfer information between elements within computer 810, such as during start-up, is typically stored in ROM 831. RAM 832 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 820. By way of example, and not limitation,
(51) The computer 810 may also include other removable/non-removable, volatile/nonvolatile computer storage media. By way of example only,
(52) The drives and their associated computer storage media, described above and illustrated in
(53) The computer 810 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 880. The remote computer 880 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 810, although only a memory storage device 881 has been illustrated in
(54) When used in a LAN networking environment, the computer 810 is connected to the LAN 871 through a network interface or adapter 870. When used in a WAN networking environment, the computer 810 typically includes a modem 872 or other means for establishing communications over the WAN 873, such as the Internet. The modem 872, which may be internal or external, may be connected to the system bus 821 via the user input interface 860 or other appropriate mechanism. A wireless networking component 874 such as comprising an interface and antenna may be coupled through a suitable device such as an access point or peer computer to a WAN or LAN. In a networked environment, program modules depicted relative to the computer 810, or portions thereof, may be stored in the remote memory storage device. By way of example, and not limitation,
(55) An auxiliary subsystem 899 (e.g., for auxiliary display of content) may be connected via the user interface 860 to allow data such as program content, system status and event notifications to be provided to the user, even if the main portions of the computer system are in a low power state. The auxiliary subsystem 899 may be connected to the modem 872 and/or network interface 870 to allow communication between these systems while the main processing unit 820 is in a low power state.
(56) Alternatively, or in addition, the functionally described herein can be performed, at least in part, by one or more hardware logic components. For example, and without limitation, illustrative types of hardware logic components that can be used include Field-programmable Gate Arrays (FPGAs), Application-specific Integrated Circuits (ASICs), Application-specific Standard Products (ASSPs), System on chip systems (SOCs), Complex Programmable Logic Devices (CPLDs), etc.
CONCLUSION
(57) While the invention is susceptible to various modifications and alternative constructions, certain illustrated embodiments thereof are shown in the drawings and have been described above in detail. It should be understood, however, that there is no intention to limit the invention to the specific forms disclosed, but on the contrary, the intention is to cover all modifications, alternative constructions, and equivalents falling within the spirit and scope of the invention.