雨果巴拉:行业北极星Vision Pro过度设计不适合市场

Microsoft Patent | Depth Map With Structured And Flood Light

Patent: Depth Map With Structured And Flood Light

Publication Number: 20200126243

Publication Date: 20200423

Applicants: Microsoft

Abstract

A method including receiving an image of a scene illuminated by both a predetermined structured light pattern and a flood fill illumination, generating an active brightness image of the scene based on the received image of the scene including detecting a plurality of dots of the predetermined structured light pattern, and removing the plurality of dots of the predetermined structured light pattern from the active brightness image, and generating a depth map of the scene based on the received image and the active brightness image.

CROSS REFERENCE TO RELATED APPLICATIONS

[0001] This application is a continuation of U.S. patent application Ser. No. 15/683,670, filed Aug. 22, 2017, the entirety of which is hereby incorporated herein by reference for all purposes.

BACKGROUND

[0002] Conventional structured light depth sensing systems typically project random dot patterns that provide sufficient texture to enable stereo matching between the camera image and the prerecorded dot pattern image.

SUMMARY

[0003] A method is provided, including receiving an image of a scene illuminated by both a predetermined structured light pattern and a flood fill illumination, generating an active brightness image of the scene based on the received image of the scene including detecting a plurality of dots of the predetermined structured light pattern, and removing the plurality of dots of the predetermined structured light pattern from the active brightness image, and generating a depth map of the scene based on the received image and the active brightness image.

[0004] This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. Furthermore, the claimed subject matter is not limited to implementations that solve any or all disadvantages noted in any part of this disclosure.

BRIEF DESCRIPTION OF THE DRAWINGS

[0005] FIG. 1 shows a side perspective view of a computing device in the form of a head mounted display (HMD) device, according to one embodiment of the present disclosure.

[0006] FIG. 2A shows an example scene being captured by the depth sensor of the computing device of FIG. 1.

[0007] FIG. 2B shows an example image of the example scene illuminated by a predetermined structured light pattern captured by the depth sensor of the computing device of FIG. 1.

[0008] FIG. 2C shows an example depth calculation using the depth sensor of the computing device of FIG. 1.

[0009] FIG. 3 shows example regions of interest for images captured by the depth sensor of the computing device of FIG. 1.

[0010] FIG. 4A shows another example image of the example scene illuminated by a flood fill illumination captured by the depth sensor of the computing device of FIG. 1.

[0011] FIG. 4B shows another example image of the example scene illuminated by both a predetermined structured light pattern and a flood fill illumination captured by the depth sensor of the computing device of FIG. 1.

[0012] FIG. 5 shows an example structured light illuminator of the depth sensor of the computing device of FIG. 1.

[0013] FIG. 6 shows an example flood fill light illuminator of the depth sensor of the computing device of FIG. 1.

[0014] FIG. 7 shows an example depth sensor of the computing device of FIG. 1.

[0015] FIG. 8 shows an example hybrid light illuminator of the depth sensor of the computing device of FIG. 1.

[0016] FIG. 9 shows another example depth sensor of the computing device of FIG. 1.

[0017] FIG. 10 shows an example method for generating a depth map using the depth sensor of the computing device of FIG. 1.

[0018] FIG. 11 shows another example method for generating a depth map using the depth sensor of the computing device of FIG. 1.

[0019] FIG. 12A shows an example image captured by the camera 24 of the depth sensor of the computing device of FIG. 1.

[0020] FIG. 12B shows an example active brightness image generated based on the image captured by the camera 24 of the depth sensor of the computing device of FIG. 1.

[0021] FIG. 13 shows an example active brightness image generated based on the image captured by the camera 24 of the depth sensor of the computing device of FIG. 1.

[0022] FIG. 14 continues the example method of FIG. 11.

[0023] FIG. 15A shows an example image captured by the camera 24 of the depth sensor of the computing device of FIG. 1.

[0024] FIG. 15B shows an example prerecorded image for the predetermined structured light pattern for the depth sensor of the computing device of FIG. 1.

[0025] FIG. 16A shows an example depth map for the example image captured by the depth sensor of the computing device of FIG. 1.

[0026] FIG. 16B shows an example support weighting for the example image captured by the depth sensor of the computing device of FIG. 1.

[0027] FIG. 17 shows an example depth map with accurate depth boundaries for the example image captured by the depth sensor of the computing device of FIG. 1.

[0028] FIG. 18 continues the example method of FIG. 11.

[0029] FIG. 19A shows an example image captured by the camera 24 of the depth sensor of the computing device of FIG. 1.

[0030] FIG. 19B shows an example prerecorded image for the predetermined structured light pattern for the depth sensor of the computing device of FIG. 1.

[0031] FIG. 20 shows an example sparse depth map for the example image captured by the depth sensor of the computing device of FIG. 1.

[0032] FIG. 21 shows an example computing system according to an embodiment of the present description.

DETAILED DESCRIPTION

[0033] As discussed above, conventional structured light depth sensing systems typically project dot patterns that provide sufficient texture to enable stereo matching between the camera image and the prerecorded dot pattern image. However, in these structured light depth sensing systems, pixels that do not observe a dot do not receive any illumination. As a consequence of this missing input, current dot-based structured light depth maps typically have inaccurately reconstructed depth boundaries. The systems and methods described herein have been devised to address these challenges.

[0034] FIG. 1 illustrates a computing device 10 in the form of a head mounted display (HMD) device 12. The HMD device 12 may be worn by a user according to an example of the present disclosure. In other examples, the computing device 10 may take other suitable forms, such as, for example, a desktop computing device, a gaming console, a laptop, a wrist mounted computing device, or a mobile computing device.

[0035] In the example of FIG. 1, the HMD device 12 includes a frame 14 that wraps around the head of the user to position a display device 16 close to the user’s eyes. The frame supports additional components of the HMD device 10, such as, for example, a processor 18 and a depth sensor 20. The depth sensor 20 may be configured to generate depth maps of a physical environment in front of the depths sensor 20 of the HMD device 12.

[0036] In one example, the depth sensor 20 includes one or more illuminators 22 and one or more cameras 24. The processor 18 includes logic and associated computer memory configured to provide image signals to the display device 16, to receive images from the one or more camera 24, and to enact various control processes described herein. For example, the processor 18 may include a logic processor and the computing device 10 may include volatile memory and non-volatile storage, as discussed in more detail below with respect to the example computing system 100 of FIG. 21, in communication with the display device 16 and the depth sensor 20.

[0037] As illustrated in FIG. 1, the depth sensor 20 includes an illuminator 22 configured to emit both a predetermined structured light pattern 26 and a flood fill illumination 28 on a scene, such as, for example, a physical environment in front of the HMD device 12 worn by a user. In the illustrated example, the emitted predetermined structured light pattern 26 is a dot based speckle pattern. The dots of a typical structured light pattern are typically small points of light and may, for example, have a diameter of one or two pixels when impinging on a light sensor of a camera 24 within the depth sensor 20 after being reflected back to the depth sensor 20. Additionally, the typical structured light pattern may include one of these dots for every 25 camera pixels. However, it will be appreciated that any suitable size of dot and dot to camera pixel density in the predetermined structured light pattern 26 may be utilized to achieve a suitable performance and accuracy, such as, dot diameters of 2, 3, or 4 pixels, and dot densities of 1 dot for every 4, 9, 16, 25, or 36 camera pixels. Additionally, as illustrated in FIG. 1, the dots of the structured light pattern are typically not placed in a uniform grid pattern. The particular pattern of dots for the predetermined structured light pattern may be generated such that each region of interest (i.e., block) encompassing one or more dots will be detectable different than each other region of interest (i.e., block). The textured provided by the pattern of dots may be used by the processor 18 to perform stereo matching between regions of interest in the predetermined structured light pattern emitted by the illuminator 22 and corresponding regions of interest in the image captured by the camera 24. It will be appreciated that the predetermined structured light pattern 26 described above and illustrated in FIG. 1 is merely exemplary, and that the predetermined structured light pattern 26 may take other suitable forms.

[0038] As further illustrated in FIG. 1, the emitted flood fill illumination 28 is a diffuse illumination that typically has a uniform intensity when emitted. Additionally, as illustrated in FIG. 1, the flood fill illumination 28 is emitted with a lower intensity of light compared to the emitted predetermined structured light pattern 26. In one example, the flood fill illumination 28 may be emitted with an a tenth, an eighth, a quarter, or half the intensity as the predetermined structured light pattern 26. However, it will be appreciated that other ratios of intensity between the flood fill illumination 28 and the predetermined structured light pattern 26 may be utilized by the illuminator 22.

[0039] The depth sensor 20 further includes a camera 24 configured to capture an image of the scene illuminated by the predetermined structured light pattern 26 and the flood fill illumination 28. In one example, the illuminator 22 is configured to emit both the predetermined structured light pattern 26 and the flood fill illumination 28 concurrently, and the camera 24 is configured to capture an image of the scene that is concurrently being illuminated by both the predetermined structured light pattern 26 and the flood fill illumination 28. In another example, the illuminator 22 is configured to sequentially emit the predetermined structured light pattern 26 and the flood fill illumination 28, and the camera 24 is configured to captured at least two images, one image of the scene illuminated by the predetermined structured light pattern 26 and a second image of the scene illuminated by the flood fill illumination 28.

[0040] The one or more images of the scene captured by the camera 24 are received by the processor 18 that is configured to generate a depth map for the scene based on the image including both the predetermined structured light pattern 26 and the flood fill illumination 28. The processor 18 may be configured to generate the depth map for the scene using a suitable structured light depth map method, such as, for example, a block or patch matching algorithm. However, it will be appreciated that any suitable structured light depth map algorithm may be utilized to generate a depth map using the images captured by camera 24 of the depth sensor 20.

[0041] FIG. 2A illustrates an example scene 30 including a background depth 32 and a foreground depth 34. In this example, the foreground depth 34 is a rectangular object that is closer to the HMD device 12 than the background depth 32, which, for example, may be a wall, table, another object, etc. As discussed previously, the illuminator 22 of the depth sensor 20 projects a predetermined structured light pattern 26 onto the example scene 30, and the camera 24 captures an image of the example scene 30 illuminated by the predetermined structured light pattern 26.

[0042] FIG. 2B illustrates an example image 36A of the example scene 30 illuminated by only the predetermined structured light pattern 26. In FIG. 2B, the size of the dots are exaggerated for ease of illustration. As show, the example image 36A includes an imaged predetermined structured light pattern 26L, which is the emitted predetermined structured light pattern 26 reflected off the example scene 30 and received by the camera 24. Due to the camera 24 being spaced away from the illuminator 22, the imaged predetermined structured light pattern 26L includes dots that have different imaged locations compared to corresponding dots in the emitted predetermined structured light pattern 26. The binocular disparity between corresponding dots is larger for dots reflected off foreground objects that are closer to the depth sensor 20, and smaller for dots reflected off background objects that are farther away from the depth sensor 20.

[0043] As illustrated in FIG. 2C, in one structured light depth map algorithm, the processor 18 may be configured to calculate depths in scene based on binocular disparities between dots of the emitted structured light pattern 26 emitted from the location of the illuminator 22 and the imaged structured light pattern 26L viewed from the location of the camera 24. As discussed previously, the dot-based predetermined structured light pattern emitted by the illuminator 22 casts rich texture onto the scene, which may be used by the processor 18 to perform stereo matching between regions of interest. In one example structured light depth map algorithm, the reference image of a stereo pair is the image of the scene captured by the camera 24. The second image of the stereo pair is a virtual one that shows what the illuminator 22 would see if it was a camera. It will be appreciated that this virtual image remains constant regardless of the scene content and can be prerecorded. That is, the virtual image is the predetermined structured light pattern emitted from the illuminator 22. In one example algorithm, for each pixel P of the reference view imaged by the camera 24, the processor 18 is configured to determine a corresponding pixel P’ in the second view of a virtual camera at the location of the illuminator 22 via a suitable stereo matching algorithm. The processor 18 calculates the corresponding 3D point P1 by intersecting the ray F1P, where F1 is the focal point of the camera 24, with the ray F2P, where F2 is the focal point of the virtual camera at the location of the illuminator 24.

[0044] In one example, the processor 18 performs stereo matching for regions of interest. FIG. 3 illustrates sixteen regions of interest including a first region of interest R1. As shown, the first region of interest R1 includes an arrangement of three dots D1, D2, and D3 having particular locations in the first region of interest R1 and particular distances from each other dot in the first region of interest R1. As shown, the particular arrangement of dots in the first region of interest R1 is different than the arrangement of dots in each other region of interest. In this manner, the texture provided by the predetermined structured light pattern 26 provides each region of interest with different and detectable arrangements of dots. Thus, in one example, the processor 18 may be configured to perform stereo matching between regions of interest in the reference image captured by the camera 24 and regions of interest in the prerecorded virtual image based on the detected arrangements of dots within the regions of interest. After determining corresponding regions of interest, the processor 18 may be configured to calculate a depth value for each region of interest in the reference image according to the method described above with reference to FIG. 2C. It will be appreciated that FIG. 3 depicts sixteen regions of interest for illustrative purposes, and that typically the processor 18 may be configured to process any suitable number of regions of interest. Typically, the processor 18 may process hundreds of regions of interest when generating a depth map for each reference image captured by the camera 24. For example, each region of interest may be a 5 by 5 pixel region. In another example, each region of interest may be a 25 by 25 pixel region. The size of the region of interest may be selected to achieve a suitable balance between performance and accuracy of the depth map.

[0045] Turning back to FIG. 2B, areas between the dots of the predetermined structured light pattern 26 typically do not receive illumination from the emitted predetermined structured light pattern 26. Thus, accurate depth data for edges of the objects in the scene that lie between dots of the predetermined structured light pattern 26 may be difficult to calculate based on only the predetermined structured light pattern 26.

[0046] As discussed previously, the illuminator 22 is further configured to emit flood fill illumination 28 onto the scene. FIG. 4A illustrates an example image 36B of the same example scene 30 of FIG. 2A illuminated by the flood fill illumination 28. The flood fill illumination 28 projects diffuse illumination of substantially uniform intensity across the area of the scene imaged by the camera 24. However, due to light decay, the intensity of the light imaged by the camera 24 will be different depending upon how close or far away an object in the scene is located from the camera 24. Light intensity of the flood fill illumination 28 reflected off closer objects will be imaged with a higher intensity compared to light intensity of the flood fill illumination 28 reflected off more distant objects in the scene. Thus, as shown in the example image 36B, the object in the scene for the foreground depth 34 is imaged with a higher intensity than the background depth 32. Additionally, as the flood fill illumination 28 projects a uniform illumination, each pixel of the image captured by the camera 24 will receive light data from the flood fill illumination 28. Thus, a large depth disparity between two pixels in the example image 36B indicates the edge of an object in the example scene 30. In this manner, the processor 18 may be configured to detect depth disparities in the image based on light intensity disparities, and detect edges of objects in the scene based on the detected depth disparities. However, while edges in the scene may be detected using the flood fill illumination 28, it will be appreciated that calculating absolute depth values of objects in the scene based on only the flood fill illumination 28 may be difficult or resource intensive.

[0047] In one example, the camera 24 is configured to capture a first image of the scene illuminated by the predetermined structured light pattern 26, and a second image of the scene illuminated by the flood fill illumination 28. In this example, the processor 18 may be configured to generate a depth map of the scene by performing stereo matching with the first image and using the second image as a guidance image for accurate edge reconstruction. However, it will be appreciated that due to a time difference between when the first image and the second image are captured, a perspective shift may occur from the user wearing the HMD device 12 moving their head. This perspective shift may reduce the accuracy of the depth map generated for the scene.

[0048] Thus, in another example the camera 24 is configured to capture an image of the scene concurrently illuminated by both the predetermined structured light pattern 26 and the flood fill illumination 28. FIG. 4B illustrates an example image 36C of the example scene 30 illuminated by both the predetermined structured light pattern 26 and the flood fill illumination 28. As shown, the flood fill illumination 28 is emitted with a lower intensity than the predetermined structured light pattern 26. In this manner, each pixel in the example image 26C receives at least some illumination. As discussed previously, the flood fill illumination 28 enables the processor 18 to determine depth discontinuities/disparities in the image to detect edges, and the predetermined structured light pattern 26 enables the processor 18 to calculate absolute depth values for each region of interest in the image.

[0049] In one example, the predetermined structured light pattern 26 and the flood fill illumination 28 are emitted as infrared light. In this example, the illuminator 22 may be configured to emit infrared light in a predetermined band of infrared light. For example, the illuminator 22 may include diode lasers configured to emit light in the infrared spectrum. In particular the diode lasers may be configured to emit light in a small predetermined band of the infrared light spectrum. Further in this example, the camera 24 may be configured to be sensitive to infrared light in the predetermined band. That is, the camera 24 may be configured to be responsive to infrared light having a wavelength within the predetermined band, and to be less responsive or unresponsive to any light having a wavelength outside the predetermined band. For example, the camera 24 may include a band pass filter configured to filter out light outside of the predetermined band of infrared light. In this manner, the depth sensor 20 may be configured to filter out potential noise from ambient light having wavelengths outside of the predetermined band.

[0050] Turning to FIG. 5, the illuminator 22 may include both a structured light illuminator and a flood fill light illuminator. FIG. 5 illustrates a structured light illuminator 38 including a first light emitter 40 and a diffractive optical element 42, the structured light illuminator 38 being configured to emit the predetermined structured light pattern 26. The diffractive optical element 42 is configured to generate constructive and destructive interference using diffractive grating to generate the predetermined structured light pattern from light emitted by the first light emitter 40. In one example, the diffractive optical element 42 receives light from a single mode laser diode having near diffraction limited beam quality, and outputs the predetermined structured light pattern 26. Thus, in this example, the first light emitter 40 may take the form of a single mode diode laser. Typically, these single mode diode lasers may have an output power capability of less than 200 mW. It will be appreciated that while the structured light illuminator 38 may generate other types of structured light patterns than the speckle pattern illustrated in FIG. 5. For example, the diffraction grating of the diffractive optical element 42 may be configured to generate other suitable illumination patterns, including other dot patterns, line based patterns, and other engineered patterns.

[0051] FIG. 6 illustrates a flood fill light illuminator 44 including a second light emitter 46 and a diffuser optical element 48, the flood fill light illuminator 44 being configured to emit the flood fill illumination 28. In one example, the diffuser optical element 48 is configured to homogenize and spread incident light using geometric optics, such as, for example, a micro-lens array. In another example, the diffuser optical element 48 may take the form of a uniform flat-top diffractive optical element configured to transform a single or multi-mode input beam into an output beam having a homogenized flat-top intensity. In this manner, the diffuser optical element 48 receives light from the second light emitter 46 taking the form of a high powered multi-mode diode laser, and outputs the flood fill illumination 28 onto the scene in front of the depth sensor 20. In this example, the second light emitter 46 is configured as a high power multi-mode diode laser, and may be configured to have a pulsed peak power greater than 30 W. In another example, the second light emitter 46 may take the form of a light emitting diode having a matched wavelength configured to emit the flood fill illumination 28 onto the scene. It will be appreciated that the example light emitters and optical elements described above are merely exemplary, and that any other suitable configurations may be used to generate the predetermined structured light pattern 26 and flood fill illumination 28.

[0052] FIG. 7 illustrates an example depth sensor 20 that includes the camera 24 and the illuminator 22 comprising both the structured light illuminator 38 and flood fill light illuminator 44 as described above. As shown, the illuminator 22 is spaced away from the camera 24 by a distance D. The distance D may be set according to the expected scenarios or environments for which the HMD device 12 will be used. A larger distance D will cause a larger binocular disparity between the reference image taken by the camera 24 and the virtual image that a camera at the location of the illuminator 22 would see. Thus, a larger distance D may increase the range of depths that may be detected using stereo techniques described herein. However, a larger distance D also increases the size of the depth sensor 20. Thus, the distance D may be set to achieve a suitable size and performance.

[0053] In one example, the structured light illuminator 38 and the flood fill light illuminator 44 are located at co-axial positions on the depth sensor 20 that are substantially equidistant from the camera 24. In the illustrated example, the structured light illuminator 38 and the flood fill light illuminator 44 are co-axial along the axis A, which is perpendicular to an axis of the distance D between the camera 24 and the illuminator 22. By being located at co-axial positions on the depth sensor 20 as illustrated in FIG. 7, the structured light illuminator 38 and the flood fill light illuminator 44 are substantially equidistant from the camera 24. Thus, in this example, the structured light illuminator 38 and the flood fill light illuminator 44 have substantially the same baseline distance D from the camera 24, which may simplify depth calculations performed by the processor 18.

[0054] As discussed previously, the illuminator 22 may be configured to emit the predetermined structured light pattern 26 and the flood fill illumination 28 concurrently or separately. In the example depth sensor 20 illustrated in FIG. 7, the structured light illuminator 38 and the flood fill illuminator 44 include separate light emitters, and may be controlled by the processor 18 to achieve a suitable timing, such as concurrently or separately. In one example, the structured light illuminator 38 and the flood fill light illuminator 44 are configured to emit the predetermined structured light pattern 26 and the flood fill illumination 28 at separate points in time, and the camera 24 is configured to capture a first image of the scene illuminated by the predetermined structured light pattern 26 and a second image of the scene illuminated by the flood fill illumination 28. An example first image of the scene is illustrated in FIG. 2B, which, as discussed above, is an example image 36A of the example scene 30 illuminated by only the predetermined structured light pattern 26. An example second image of the scene is illustrated in FIG. 4A, which, as discussed above, is an example image 36B of the example scene 30 illuminated by only the flood fill illumination 28. Based on the first and second images, the processor 18 may be configured to use the depth disparities detected in the second image 26B as edge guidance for the depth map generated using stereo matching algorithms with the first image 36A. Emitting light from the structured light illuminator 38 and the flood light illuminator 44 at separate points in time may potentially reduce overall output power required for illumination and reduce shot noise from the sensor. However, as discussed previously, time elapsed between capture of the first and second images may potentially introduce motion blur between the captured images due to the user of the HMD device 12 moving their head between images.

[0055] In another example, the structured light illuminator 38 and the flood fill light illuminator 44 are configured to emit the predetermined structured light pattern 26 and the flood fill illumination 28 concurrently, and the camera 24 is configured to capture an image of the scene concurrently illuminated by both the predetermined structured light pattern 26 and the flood fill illumination 28. An example image of the scene is illustrated in FIG. 4B, which, as discussed above, is an example image 36C of the example scene 30 illuminated by both the predetermined structured light pattern 26 and the flood fill illumination 28. Similarly, the processor 18 may be configured to use depth disparities detected based on the imaged flood fill illumination 28 as edge guidance for the depth map generated using stereo matching algorithms with the imaged predetermined structured light pattern 26. By capturing a single image of the scene concurrently illuminated with both the predetermined structured light pattern 26 and the flood fill illumination 28, potential motion blur may be minimized. However, concurrently emitting both the predetermined structured light pattern 26 and the flood fill illumination 28 may potentially increase the overall output power required for the depth sensor 20.

[0056] FIG. 8 illustrates a hybrid light illuminator 50 including a light emitter 52 and a hybrid diffractive optical element 54, the hybrid light illuminator 50 being configured to emit both the predetermined structured light pattern 26 and the flood fill illumination 28. As shown, the emitted light 56 from the hybrid light illuminator 50 is the superposition of the predetermined structured light pattern 26 and the flood fill illumination 28. In one example, the hybrid diffractive optical element 54 includes a first diffractive portion 54A configured for the predetermined structured light pattern 26 and a second diffractive portion 54B configured for the flood fill illumination 28. That is, the first diffractive portion 54A includes a diffraction pattern for the predetermined structured light pattern 26 and the second diffractive portion 54B includes a tophat diffraction pattern for the flood light illumination 28. Light emitted from the light emitter 52 samples both the first diffractive portion 54A and the second diffractive portion 54B while passing through the hybrid diffractive optical element 54, producing a light pattern that is the superposition of the predetermined structured light pattern 26 and the flood fill illumination 28.

[0057] In one example, a ratio between the first diffractive portion 54A and the second diffractive portion 54B may be set to achieve a suitable ratio between the intensity of the emitted predetermined structured light pattern 26 and the flood fill illumination 28. For example, the hybrid diffractive optical element 54 may include a larger portion of the first diffractive portion 54A compared to the second diffractive portion 54B, such that the hybrid light illuminator 50 emits the flood fill illumination 28 with a lower light intensity than the predetermined structured light pattern 26. In one specific example, the hybrid light illuminator 50 may be configured to emit the flood fill illumination 28 with a tenth, an eighth, a quarter, or half the light intensity of the predetermined structured light pattern. However, it will be appreciated that the hybrid light illuminator 50 may be configured to emit other suitable ratios of light intensity between the flood fill illumination 28 and the predetermined structured light pattern 26, such as, for example, a 1 to 6 ratio, a 1 to 5 ratio, etc.

[0058] FIG. 9 illustrates an example depth sensor 20 that includes a camera 24 and an illuminator 22 comprising the hybrid light illuminator 50. In one example, the predetermined structured light pattern 26 and the flood fill illumination 28 are emitted from the hybrid light illuminator 50 co-aligned. That is, both the predetermined structured light pattern 26 and the flood fill illumination 28 are emitted from the same light source, the hybrid light illuminator 50. As illustrated, the hybrid light illuminator 50 is spaced away from the camera 24 by a baseline distance D. As discussed previously, the baseline distance D may be set to achieve a suitable size and performance of the depth sensor 20.

[0059] FIG. 10 shows an example method 1000 for generating depth maps using the depth sensor 20 described herein. At step 1002, the method 1000 may include emitting both a predetermined structured light pattern 26 and a flood fill illumination 28 on a scene. As discussed above, the depth sensor 20 includes an illuminator 22 configured to emit both the predetermined structured light pattern 26 and the flood fill illumination 28. In one example, the illuminator 22 may include a structured light illuminator 38 and a flood fill light illuminator 44, which may be configured to emit the predetermined structured light pattern 26 and the flood fill illumination 28 concurrently or at separate points in time. In another example, the illuminator 22 may include a hybrid light illuminator 50 configured to coaxially emit both the predetermined structured light pattern 26 and the flood fill illumination 28 onto the scene.

[0060] At step 1004, the method 1000 may include capturing an image of the scene illuminated by the predetermined structured light pattern 26 and the flood fill illumination 28. The image of the scene is captured by the camera 24 of the sensor device 20. In one example where the illuminator 22 is configured to emit the predetermined structured light pattern 26 and the flood fill illumination 28 at separate points in time, step 1004 may include capturing a first image of the scene illuminated by the predetermined structured light pattern 26, and a second image of the scene illuminated by the flood fill illumination 28.

[0061] At step 1006, the method 1000 may include generating a depth map for the scene based on the image including both the predetermined structured light pattern 26 and the flood fill illumination 28. Several methods for generating depths maps with accurate depth boundary reconstruction based on the image of the scene including both the predetermined structured light pattern 26 and the flood fill illumination 28 are discussed below.

您可能还喜欢...