Home Technology peripherals AI Microsoft AR/VR patent shares motion correction method for ToF imaging

Microsoft AR/VR patent shares motion correction method for ToF imaging

Oct 19, 2023 am 11:25 AM

(Nweon October 19, 2023) Depth sensing systems such as time-of-flight ToF cameras can be used to generate depth images of the environment for HoloLens 2, where each pixel of the depth image represents a corresponding point in the environment distance. In ToF imaging, the distance of a point on the imaging surface in the environment depends on the length of the time interval for the light from the ToF camera to travel to that point and then return to the ToF camera sensor.

Phase-based ToF imaging is a variant of ToF imaging in which depth is calculated based on the phase shift of amplitude-modulated light reflected back from the object. Due to the movement of the ToF camera between phase data acquisitions, intra-frame phase data may be relatively shifted. For example, in a first intra-frame phase data acquisition, a pixel may sense phase data at a first modulation frequency at a location within the scene. Then during phase data acquisition within the next frame, the pixels can sense phase data at a different scene location at a second modulation frequency. The phase data at different frequencies becomes inconsistent, which can be called motion blur.

So in the patent application titled "Motion correction for time-of-flight depth imaging", Microsoft proposed a motion correction method for ToF imaging.

To do this, the active brightness AB image corresponding to the first illumination light modulation frequency is compared with the AB image corresponding to the second illumination light modulation frequency to determine the intra-frame two-dimensional translation. Intra-frame 2D translation corresponds to the estimated motion of the camera relative to the imaged scene.

Intra-frame 2D translation is then applied to the phase data to help align the phase images and form corrected phase data. Next, phase unpacking is performed on the corrected phase data to obtain an intra-frame corrected three-dimensional depth image. Intra-frame 2D translation can also be used to perform motion correction on intra-frame AB images. For example, the intra-frame corrected AB image can be obtained by averaging the corrected AB images.

In such an example, the first AB image from the first frame is compared to the second AB image from the second frame to determine the inter-frame 2D translation. Inter-frame translation is an estimate of motion between frames. An inter-frame 2D translation can then be applied to the depth image of the first frame to form an inter-frame generated depth image. The inter-frame two-dimensional translation can be determined from the intra-frame corrected AB image.

Additionally, as an estimate of camera motion, inter-frame 2D translation can be output and used in various temporal post-processing routines or software services. Therefore, the embodiments described in the patent can help achieve motion blur correction of 3D depth data using 2D intra-frame AB image data. By using 2D data, motion correction can be performed in real time with relatively efficient computational performance compared to using 3D depth data for motion blur correction.

Microsoft AR/VR patent shares motion correction method for ToF imaging

Figure 2 shows an example including a ToF camera 202. Among them, the ToF camera is a phase-based ToF depth imaging system 200 . ToF camera 202 includes a sensor array 204 including a plurality of ToF pixels 206 each configured to acquire a light sample that captures phase data, a controller 208 and an objective system 210 .

The controller 208 is configured to collect and process data from the ToF pixels 206 of the sensor array 204 to construct a depth image. Controller 208 may include executable instructions to perform denoising and/or phase unwrapping.

Depth imaging system 200 includes both a modulated light emitter 230 and an analog and/or digitally modulated electronic shutter 232 for sensor array 204 to control the integration of light through sensor array 204. Modulated light emitter 230 may be configured to emit electromagnetic radiation with any frequency detectable by ToF pixel 206 .

The modulated light may be modulated at different frequencies sequentially or simultaneously, and the sensor array 204 is configured to sample the light reflected from the modulated light emitter 230 to the surface 220 and back to the camera. Each ToF sensing pixel 206 of sensor array 204 may include one or more pixel taps to integrate reflected light signals at different time intervals and thereby determine phase shifts.

For each modulation frequency, the sensor array 204 is controlled to sample the light at multiple phase angles of the amplitude-modulated light from the light source and determine the light for each modulation frequency from the multiple light samples for the modulation frequency. phase samples. The phase samples can then be unwrapped to obtain a depth value for each pixel.

Due to the periodicity of the modulated light, the total phase measured repeats every 2π. Since n(k) cannot be directly measured by phase-based ToF pixels, the total phase and therefore the actual distance associated with the measurement are ambiguous. Therefore in phase-based ToF imaging, the distance that can be measured (blur-free range) is limited by the modulation frequency.

Two or more different modulation frequencies can be used to increase the ambiguity-free range, and then the collected phase shift data can be spread out to accurately determine distance.

Microsoft AR/VR patent shares motion correction method for ToF imaging

Figure 3 schematically illustrates example ToF image data 300 for a plurality of K modulation frequencies. Data 300 represents data that can be acquired by the depth imaging system 200 during multi-frequency frame acquisition.

In the example shown, the depth data includes an M×N array of data for each of the K modulation frequencies, resulting in an M×N grid of intra-frame depth data 302a-c, where each mesh Each pixel 304 in the grid represents a measurement value obtained at a corresponding illumination light modulation frequency K of K modulation frequencies.

The measured phase is used to calculate the depth value associated with the pixel. But as mentioned above, in phase-based ToF imaging, the distance that can be measured (blur-free range) is limited by the modulation frequency. Therefore, a set of K ≥ 2 modulation frequencies K can be used to increase the range, allowing the phase information to be unraveled to accurately determine distance.

Phase unwrapping is a method of disambiguating phase shift data and identifying correct distance values ​​by illuminating a scene with multiple amplitude-modulated lights of different frequencies, since the distance ambiguity is different for each frequency of illumination light.

But as mentioned above, if the ToF depth camera is in motion, phase unwrapping errors may occur. Since the depth data acquisition at each frequency is performed sequentially, the phase image and AB image are temporarily separated within one frame.

For example, intra-frame phase image 306a may include first frequency ƒ1 phase data acquired toward the beginning of the frame, and intra-frame phase image 306b may include second frequency ƒ2 phase data acquired toward the middle of the frame, intra-frame Phase image 306b may include third frequency ƒ3 phase data acquired toward the end of the frame.

Therefore, if the ToF depth camera moves between intra-frame acquisitions, the phase data at the three different frequencies may be shifted and misaligned. Inconsistent phase data can lead to errors in phase unrolling.

So, Microsoft proposed a solution that uses AB image data to estimate motion and determine two-dimensional translation.

The company notes that while depth data can be used, determining 2D translation based on AB image data may be more robust due to active brightness differences between objects in the scene. Intra-frame motion correction can then be performed on the phase data using 2D translation. Performing phase unwrapping on intra-frame corrected phase data can help avoid unwrapping errors due to motion blur.

Microsoft AR/VR patent shares motion correction method for ToF imaging

Figure 4 shows an example method for performing such correction on depth data prior to phase unwrapping. Method 400 may be implemented in a computing system that receives data from a ToF depth camera.

At 402, multi-frequency frame collection is performed, in which multiple intra-frame phase samples are collected by the ToF image sensor. Intra-frame phase samples are collected for each of a plurality of illumination light modulation frequencies (ƒ1, ƒ2, ƒ3) to form corresponding intra-frame depth data 404a-c. In this example, intra-frame depth data 404a is obtained first, intra-frame depth data 404b is obtained secondly, and intra-frame depth data 404c is obtained thirdly, as shown by the time arrow.

At 406, signal calibration correction is performed to obtain phase data 408 and active brightness data 410. As mentioned above, intra-frame depth data at different frequencies may be relatively shifted due to camera motion. Therefore, method 400 compares active brightness data 410 to estimate and correct for motion.

Here, the intra-frame AB image 410a is compared with the intra-frame AB image 410b to determine the first intra-frame two-dimensional translation 412. The two-dimensional translation from AB1 to AB2 can be expressed as [Δu, Δv]1,2, where Δu is the pixel displacement in the x direction and Δv is the pixel displacement in the y direction. Intra-2D translation 412 is an estimate of motion between intra-frame depth data 404a and intra-frame depth data 404b.

Next, the intra-AB image 410a is compared to the intra-AB image 410c to determine a second intra-frame 2D translation 414, denoted as [Δu, Δv] 13.

In one embodiment, intra-AB image 410b may be compared to intra-AB image 410c to determine a third intra-frame two-dimensional translation. In other examples, any suitable intra-frame AB image pairs may be compared to determine corresponding intra-frame 2D translations. Any suitable method can be used to compare the AB images and determine the translation. In one example, features are extracted from AB images, formed into feature maps, and used to compare images.

After determining the 2D translation, the phase data can be corrected using the determined translation. In the embodiment shown in FIG. 4 , intra-frame two-dimensional translation 412 is applied to the phase image 408b to correct the phase image to form a corrected phase image 420b. Similarly, intra-frame two-dimensional translation 414 is performed on the phase image 408c to correct the phase image, forming a corrected phase image 420c.

Therefore, the corrected phase data 420 represents the phase data that has been "realigned" to the phase image 408a. In one example, one or more two-dimensional translations may be applied to form a corrected phase image aligned with phase image 408b or 408c.

At 422, method 400 also includes performing phase unwrapping on the corrected phase data 420 to form a depth image 424. Since the corrected phase images 420b, 420c can be realigned with the phase image 408a, the phase unwrapping at 422 can produce relatively less unwrapping error than in the example where motion blur correction is ignored.

So, method 400 can help achieve better performance in processing depth data 404 to form depth image 424.

Additionally, the application of intra-frame 2D translation can be performed via convolution and thus can be combined with spatial and temporal filtering processes that also utilize convolution. At the same time, as an estimate of camera motion, the determined intra-frame 2D translation may be useful for various post-processing applications, such as temporal filters, trajectory estimation, dynamic region estimation or mapping. Similar techniques can be used to correct intra-AB images and form intra-corrected AB images.

Microsoft AR/VR patent shares motion correction method for ToF imaging

Referring to Figure 5, method 500 utilizes intra-frame two-dimensional translation 412, 414 to form corrected AB data 510. In one example, method 500 is performed together with method 400. In other examples, method 500 is performed separately from method 400 .

As shown in Figure 5, the intra-frame two-dimensional translation 412 is applied to the intra-frame AB image 410b to correct the image to form a corrected AB image 510b. Further, intra-frame two-dimensional translation 414 is applied to AB image 410c to form corrected AB image 510c. Together with intra-AB image 408a, corrected AB images 510b-c form corrected AB data 510.

At 518, method 500 further includes averaging the corrected AB data 510 to form an intra-frame corrected AB image 520. The intra-corrected AB image 520 may be output for further processing and/or output to a display.

Intra-frame corrected AB images can also be used to generate inter-frame depth images. In addition to correcting motion blur within depth image frames, correction can also be performed between frames.

Related Patents: Microsoft Patent | Motion correction for time-of-flight depth imaging

The Microsoft patent application titled "Motion correction for time-of-flight depth imaging" was originally submitted in March 2022 and was recently published by the US Patent and Trademark Office.

The above is the detailed content of Microsoft AR/VR patent shares motion correction method for ToF imaging. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

Getting Started With Meta Llama 3.2 - Analytics Vidhya Getting Started With Meta Llama 3.2 - Analytics Vidhya Apr 11, 2025 pm 12:04 PM

Meta's Llama 3.2: A Leap Forward in Multimodal and Mobile AI Meta recently unveiled Llama 3.2, a significant advancement in AI featuring powerful vision capabilities and lightweight text models optimized for mobile devices. Building on the success o

10 Generative AI Coding Extensions in VS Code You Must Explore 10 Generative AI Coding Extensions in VS Code You Must Explore Apr 13, 2025 am 01:14 AM

Hey there, Coding ninja! What coding-related tasks do you have planned for the day? Before you dive further into this blog, I want you to think about all your coding-related woes—better list those down. Done? – Let&#8217

AV Bytes: Meta's Llama 3.2, Google's Gemini 1.5, and More AV Bytes: Meta's Llama 3.2, Google's Gemini 1.5, and More Apr 11, 2025 pm 12:01 PM

This week's AI landscape: A whirlwind of advancements, ethical considerations, and regulatory debates. Major players like OpenAI, Google, Meta, and Microsoft have unleashed a torrent of updates, from groundbreaking new models to crucial shifts in le

Selling AI Strategy To Employees: Shopify CEO's Manifesto Selling AI Strategy To Employees: Shopify CEO's Manifesto Apr 10, 2025 am 11:19 AM

Shopify CEO Tobi Lütke's recent memo boldly declares AI proficiency a fundamental expectation for every employee, marking a significant cultural shift within the company. This isn't a fleeting trend; it's a new operational paradigm integrated into p

A Comprehensive Guide to Vision Language Models (VLMs) A Comprehensive Guide to Vision Language Models (VLMs) Apr 12, 2025 am 11:58 AM

Introduction Imagine walking through an art gallery, surrounded by vivid paintings and sculptures. Now, what if you could ask each piece a question and get a meaningful answer? You might ask, “What story are you telling?

GPT-4o vs OpenAI o1: Is the New OpenAI Model Worth the Hype? GPT-4o vs OpenAI o1: Is the New OpenAI Model Worth the Hype? Apr 13, 2025 am 10:18 AM

Introduction OpenAI has released its new model based on the much-anticipated “strawberry” architecture. This innovative model, known as o1, enhances reasoning capabilities, allowing it to think through problems mor

How to Add a Column in SQL? - Analytics Vidhya How to Add a Column in SQL? - Analytics Vidhya Apr 17, 2025 am 11:43 AM

SQL's ALTER TABLE Statement: Dynamically Adding Columns to Your Database In data management, SQL's adaptability is crucial. Need to adjust your database structure on the fly? The ALTER TABLE statement is your solution. This guide details adding colu

Reading The AI Index 2025: Is AI Your Friend, Foe, Or Co-Pilot? Reading The AI Index 2025: Is AI Your Friend, Foe, Or Co-Pilot? Apr 11, 2025 pm 12:13 PM

The 2025 Artificial Intelligence Index Report released by the Stanford University Institute for Human-Oriented Artificial Intelligence provides a good overview of the ongoing artificial intelligence revolution. Let’s interpret it in four simple concepts: cognition (understand what is happening), appreciation (seeing benefits), acceptance (face challenges), and responsibility (find our responsibilities). Cognition: Artificial intelligence is everywhere and is developing rapidly We need to be keenly aware of how quickly artificial intelligence is developing and spreading. Artificial intelligence systems are constantly improving, achieving excellent results in math and complex thinking tests, and just a year ago they failed miserably in these tests. Imagine AI solving complex coding problems or graduate-level scientific problems – since 2023

See all articles