US12445588B2 - Adaptive displays to address the vergence-accommodation conflict - Google Patents
Adaptive displays to address the vergence-accommodation conflictInfo
- Publication number
- US12445588B2 US12445588B2 US18/204,599 US202318204599A US12445588B2 US 12445588 B2 US12445588 B2 US 12445588B2 US 202318204599 A US202318204599 A US 202318204599A US 12445588 B2 US12445588 B2 US 12445588B2
- Authority
- US
- United States
- Prior art keywords
- microdisplay
- segment
- content
- microdisplays
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/388—Volumetric displays, i.e. systems where the image is built up from picture elements distributed through a volume
- H04N13/393—Volumetric displays, i.e. systems where the image is built up from picture elements distributed through a volume the volume being generated by a moving, e.g. vibrating or rotating, surface
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/296—Synchronisation thereof; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/344—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/365—Image reproducers using digital micromirror devices [DMD]
Definitions
- the present disclosure relates to stereoscopy, and in particular to techniques for improving stereo image displays.
- VR systems may fully immerse (e.g., giving the user a sense of being in an environment) or partially immerse (e.g., giving the user the sense of looking at an environment) users in a three-dimensional (3D), computer-generated environment.
- AR systems may provide a modified version of reality, such as enhanced information overlaid over real-world objects.
- MR systems map and merge virtual objects, which are often interactive, to the real world.
- Such XR systems may utilize wearables (e.g., head-mounted devices, smart glasses, etc.) comprising a stereoscopic display to generate images that convey 3D depth to a user.
- Stereoscopic displays utilize stereoscopy, which is a technique for creating or enhancing the illusion of depth in an image by means of stereopsis for binocular vision.
- stereoscopy is a technique for creating or enhancing the illusion of depth in an image by means of stereopsis for binocular vision.
- One of the major limitations of stereoscopic displays is that the stereo images are displayed on a fixed focal plane.
- the muscles in the eye naturally turn the optical axis of an eye to point towards an object of interest allowing the left eye and the right eye to verge on the object of interest.
- objects are located at different distances from the human visual system.
- the human visual system uses the differing distances to perceive depth.
- most immersive content is displayed on displays that are on a fixed viewing plane relative to the human visual system, causing a vergence-accommodation conflict.
- the vergence-accommodation conflict can result in eye strain, headache, nausea, reduced reaction time, reduced vision clarity, double vision, and similar such ailments.
- a device may receive a content item (e.g., a piece of 3D content).
- the 3D content may be an image or plurality of images and may correspond to a movie, television show, video game, and/or any other type of 3D content.
- the device may determine that the 3D content comprises a first object.
- the 3D content may depict a scene, where the first object is located on or near a first plane and a second object is located on or near a second plane.
- a portion of the first object may be located on the first plane, but the entire first object may be associated with the first plane as an approximation.
- no portion of the first object may be located on the first plane, but the entire object may be associated with the first plane as an approximation because the first object is near the first plane.
- all portions of the first object may be located on the first plane.
- the device may manipulate one or more portions of a display.
- the device may change a location of a first microdisplay from a first position to a second position based on the first object being depicted on the first plane within the 3D content.
- the device may then display the 3D content on the display.
- the portion of the display being changed from the first position to the second position provides depth for rendering the object at a suitable accommodation distance and reduces or eliminates the vergence-accommodation conflict.
- the device may repeat this process for subsequent images of the 3D content. For example, a subsequent image of the 3D content may depict the first object on a third plane within the scene.
- the device may manipulate the one or more portions of the display. For example, the device may change the location of the first microdisplay from the second position to a third position based on the first object being depicted on the third plane within the 3D content.
- the device may update the positions of the one or more microdisplays according to objects depicted in the 3D content allowing the device to display 3D videos with reduced or eliminated vergence-accommodation conflict.
- each portion of the display corresponds to one or more microdisplays of a plurality of microdisplays, wherein one or more microdisplays of the plurality of microdisplays are adjustable.
- a microdisplay refers to a small display (e.g., a display that is less than five centimeters diagonal).
- the one or more microdisplays may be adjustable due to the device utilizing a micro-electromechanical system (MEMS).
- MEMS micro-electromechanical system
- a first microdisplay may be attached to one or more telescopic support rods. The device may use the one or more telescopic support rods to adjust the position of the first microdisplay to display 3D content.
- each portion of the device corresponds to a portion of a flexible display, wherein one or more portions are adjustable.
- a first portion of the display may contact one or more micropillars.
- the device may use the one or more micropillars to adjust the position of the first portion of the display to display 3D content.
- a first portion of the display may contact one or more fluid chambers.
- the device may use the one or more fluid chambers to adjust the position of the first portion of the display to display 3D content.
- FIGS. 1 A and 1 B show illustrative diagrams of a device for manipulating a stereoscopic display so that portions of the stereoscopic display are located at varying planes, in accordance with embodiments of the disclosure.
- FIGS. 2 A and 2 B show illustrative diagrams of another device for manipulating a stereoscopic display so that portions of the stereoscopic display are located at varying planes, in accordance with embodiments of the disclosure.
- FIGS. 3 A- 3 C show illustrative diagram of another device for manipulating a stereoscopic display so that portions of the stereoscopic display are located at varying planes, in accordance with embodiments of the disclosure.
- FIG. 4 shows an illustrative diagram of parameters used to determine positions for one or more portions of a display, in accordance with embodiments of the disclosure.
- FIGS. 5 A and 5 B show illustrative diagrams of displays comprising multiple portions, in accordance with embodiments of the disclosure.
- FIGS. 6 A and 6 B show illustrative diagrams of a mechanism for manipulating a display so that portions of the display are located at varying planes, in accordance with embodiments of the disclosure.
- FIGS. 7 A- 7 D show illustrative diagrams of another mechanism for manipulating a display so that portions of the display are located at varying planes, in accordance with embodiments of the disclosure.
- FIGS. 8 A and 8 B show illustrative diagrams of a lens used with a device for manipulating a stereoscopic display so that portions of the stereoscopic display are located at varying planes, in accordance with embodiments of the disclosure.
- FIGS. 9 A and 9 B show illustrative diagrams of another lens used with a device for manipulating a stereoscopic display so that portions of the stereoscopic display are located at varying planes, in accordance with embodiments of the disclosure.
- FIG. 10 shows an illustrative block diagram of a media system, in accordance with embodiments of the disclosure.
- FIG. 11 shows an illustrative block diagram of a user equipment device system, in accordance with some embodiments of the disclosure.
- FIG. 12 is an illustrative flowchart of a process for manipulating a display so that portions of the display are located at varying planes, in accordance with embodiments of the disclosure.
- FIG. 13 is another illustrative flowchart of a process for manipulating a display so that portions of the display are located at varying planes, in accordance with embodiments of the disclosure.
- FIGS. 1 A and 1 B show illustrative diagrams of a device 100 comprising a first lens 102 a and a second lens 102 b .
- the device 100 is a head-mounted device used to display 3D content.
- the device 100 may also comprise a first microdisplay 104 a , a second microdisplay 104 b , a third microdisplay 104 c , and a fourth microdisplay 104 d .
- a user views one or more microdisplays by looking through the first lens 102 a and/or the second lens 102 b .
- one or more microdisplays correspond to a lens.
- the device 100 also comprises a first member 106 a , a second member 106 b , a third member 106 c , and a fourth member 106 d coupled to a housing 108 .
- the members connect one or more microdisplays to the housing 108 .
- the first member 106 a may connect the first microdisplay 104 a to the housing 108 and the second member 106 b may connect the second microdisplay 104 b to the housing 108 .
- the device 100 may comprise 500 microdisplays.
- each microdisplay is connected to the housing 108 with one member, any number of members may be used.
- the first microdisplay 104 a may be connected to the housing 108 by four members. In some embodiments, not all components of the device 100 are shown to avoid overcomplicating the drawing.
- the device 100 receives a piece of 3D content.
- the device 100 may receive a 3D movie from a server.
- the piece of 3D content is an image or plurality of images corresponding to a movie, television show, video game, and/or any other type of 3D content.
- the piece of 3D content is in a first format (e.g., a mesh file format, light field file format, etc.).
- the first format comprises or defines the geometry, including depth, of one or more objects.
- the first format comprises or defines fundamental elements (e.g., polygons comprising the one or more objects in 3D space) related to the one or more objects.
- the device 100 may determine that the piece of 3D content comprises a first object.
- the piece of 3D content may depict a scene comprising a house and a cloud, where the first object (e.g., cloud) is located on a first plane and a second object (e.g., house) is located on a second plane.
- the device 100 determines that the piece of 3D content comprises the first object using metadata associated with the piece of 3D content.
- the device 100 may receive metadata associated with the piece of 3D content before/after the device 100 receives the piece of 3D content, wherein the metadata indicates that the piece of 3D content comprises 3D objects.
- the device 100 may receive metadata associated with the piece of 3D content when the device 100 receives the piece of 3D content.
- the metadata may also provide information about one or more objects displayed in the piece of 3D content.
- the metadata may indicate location information relating to the first object displayed in the piece of 3D content.
- the location information may comprise coordinates related to one or more objects displayed in the piece of 3D content, indicators specifying plane information related to the one or more object displayed in the piece of 3D content, and/or similar such information.
- the piece of 3D content comprises the metadata.
- the device 100 may manipulate one or more microdisplays in response to determining that the piece of 3D content comprises the first object. For example, the device 100 may determine that the first object is located on a first plane within the scene and a second object is located on a second plane within the scene. The device 100 may then determine one or more locations for the microdisplays according to the planes associated with the objects in the scene. For example, the device 100 may calculate a position for the second microdisplay 106 b to display the first object based on a focal length of the first lens 102 a and the first object being located on the first plane within the scene. The device 100 may then change the positions of one or more microdisplays to display the objects according to the objects being located on varying planes within the scene.
- the device 100 may change the second microdisplay 104 b from a first position (e.g., FIG. 1 A ) to a second position (e.g., FIG. 1 B ).
- the device 100 may also change the fourth microdisplay 104 d from a first position (e.g., FIG. 1 A ) to a second position (e.g., FIG. 1 B ).
- the change in positions of the second microdisplay 104 b and the fourth microdisplay 104 d may be the same.
- the change in positions of the second microdisplay 104 b and the fourth microdisplay 104 d may be different.
- the distance between the second microdisplay 104 b and the first lens 102 a may need to be closer or farther than the distance between the fourth microdisplay 104 d and the second lens 102 b to provide the stereoscopic display of the piece of 3D content.
- the device 100 changes the position of the second microdisplay 104 b using the second member 106 b and changes the position of the fourth microdisplay 104 d using the fourth member 106 d .
- the second member 106 b and the fourth member 106 d may be telescopic support rods that can be used to change the positions of the second microdisplay 104 b and the fourth microdisplay 104 d .
- the second member 106 b and the fourth member 106 d may be fluid chambers that can be used to change the positions of the second microdisplay 104 b and the fourth microdisplay 104 d .
- the device 100 utilizes a MEMS comprising the second member 106 b and the fourth member 106 d to change the positions of the second microdisplay 104 b and the fourth microdisplay 104 d .
- all the members ( 106 a - 106 d ) are the same type of members.
- all the members may be telescopic support rods.
- the members ( 106 a - 106 d ) are not the same type of members.
- the first member 106 a and the second member 106 b may be fluid chambers and the third member 106 c and the fourth member 106 d may be telescopic support rods.
- the device 100 may then display the first object using the second microdisplay 104 b and the fourth microdisplay 104 d .
- the device 100 displays the second object using the first microdisplay 104 a and the third microdisplay 104 c .
- a user may view the first microdisplay 104 a and the second microdisplay 104 b using their left eye and view the third microdisplay 104 c and the fourth microdisplay 104 d using their right eye.
- the device 100 displaying the first object using microdisplays e.g., the second microdisplay 104 b and the fourth microdisplay 104 d ) on different planes than microdisplays (e.g., the first microdisplay 104 a and the third microdisplay 104 c ) displaying the second object provides depth and reduces or eliminates the vergence-accommodation conflict.
- the microdisplays are arranged so that the user is unable to observe gaps between the microdisplays.
- the device 100 further comprises fiber optic fabric.
- the fiber optic fabric may be connected between microdisplays.
- a first fiber optic fabric may be connected between the first microdisplay 104 a and the second microdisplay 104 b .
- fiber optic fabric ensures that there are no observable gaps between microdisplays.
- boundaries of the microdisplays are propagated to the fiber optic fabric.
- a fiber optic fabric may be connected to a first edge of the first microdisplay 104 a and a first edge of the second microdisplay 104 b .
- the first edge of the first microdisplay 104 a may display one or more colors.
- the one or more colors may be propagated onto the fiber optic fabric.
- the fiber optic fabric displaying the one or more colors of the first edge of the first microdisplay 104 a provides continuity between the first microdisplay 104 a and the second microdisplay 104 b.
- FIGS. 2 A and 2 B show illustrative diagrams of a device 200 comprising a first lens 202 a and a second lens 202 b .
- the device 200 is a head-mounted device used to display 3D content.
- the device 200 may also comprise a first microdisplay 204 a and a second microdisplay 204 b .
- a user views one or more microdisplays by looking through the first lens 202 a and/or the second lens 202 b .
- one or more microdisplays correspond to a lens.
- the first microdisplay 204 a is only visible through the first lens 202 a and the second microdisplay 204 b may only be visible through the second lens 202 d .
- the first microdisplays 204 a and the second microdisplays 204 b are flexible.
- the first microdisplay 204 a and the second microdisplay 204 b are shown as two separate displays, the first microdisplay 204 a and the second microdisplay 204 b may be two portions of one single display.
- the device 200 also comprises a first member 206 a , a second member 206 b , a third member 206 c , and a fourth member 206 d coupled to a housing 208 .
- the members connect the microdisplays to the housing 208 .
- the first member 206 a and the second member 206 b may connect the first microdisplay 204 a to the housing 208 and the third member 206 c and the fourth member 206 d may connect the second microdisplay 204 b to the housing 208 .
- any number of microdisplays may be housed within the device 200 .
- each microdisplay is connected to the housing 208 with two members, any number of members may be used. In some embodiments, not all components of the device 200 are shown to avoid overcomplicating the drawing.
- the device 200 receives a piece of 3D content and determines that the piece of 3D content comprises a first object. In some embodiments, the device 200 receives the piece of 3D content and/or determines that the piece of 3D content comprises the first object using the same or similar methodologies as described above in FIGS. 1 A and 1 B .
- the microdisplays may comprise one or more portions.
- the device 200 may comprise a first portion 210 a of the first microdisplay 204 a , a second portion 210 b of the first microdisplay 204 a , and a third portion 210 c of the first microdisplay 204 a .
- the device 200 may also comprise a first portion 212 a of the second microdisplay 204 b , a second portion 212 b of the second microdisplay 204 b , and a third portion 212 c of the second microdisplay 204 b.
- the device 200 may manipulate one or more microdisplays and/or portions of one or more microdisplays in response to determining that the piece of 3D content comprises the first object. For example, the device 200 may determine that the first object may be located on a first plane within the scene and a second object is located on a second plane within the scene. The device 200 may then determine one or more locations for portions of the microdisplays according to the planes associated with the objects in the scene. For example, the device 200 may calculate a position for the third portion 210 c of the first microdisplay 204 a to display the first object based on a focal length of the first lens 202 a and the first object being located on the first plane within the scene.
- the device 200 may then change the positions of one or more portions of the microdisplays to display one or more objects according to the one or more objects being located on varying planes within the scene. For example, the device 200 may change the third portion 210 c of the first microdisplay 204 a from a first position (e.g., FIG. 2 A ) to a second position (e.g., FIG. 2 B ). The device 200 may also change the third portion 212 c of the second microdisplay 204 b from a first position (e.g., FIG. 2 A ) to a second position (e.g., FIG. 2 B ).
- the change in positions of the third portion 210 c of the first microdisplay 204 a and the third portion 212 c of the second microdisplay 204 b may be the same. In some embodiments, the change in positions of the third portion 210 c of the first microdisplay 204 a and the third portion 212 c of the second microdisplay 204 b may be different. For example, the distance between the third portion 210 c of the first microdisplay 204 a and the first lens 202 a may need to be closer or farther than the distance between the third portion 212 c of the second microdisplay 204 b and the second lens 202 b to provide the stereoscopic display of the 3D content.
- the device 200 changes the position of the third portion 210 c of the first microdisplay 204 a using the second member 206 b and changes the position of the third portion 212 c of the second microdisplay 204 b using the fourth member 206 d .
- the members are telescopic support rods, fluid chambers, micropillars, and/or similar such components able to change the positions of one or more portions of a microdisplay.
- the device 200 utilizes a MEMS comprising one or more members to change the positions of the third portion 210 c of the first microdisplay 204 a and the third portion 212 c of the second microdisplay 204 b .
- all the members ( 206 a - 206 d ) are the same type of members. In some embodiments, the members ( 206 a - 206 d ) are not the same type of members.
- the device 200 may then display the piece of 3D content using the first microdisplay 204 a and the second microdisplay 204 b .
- the device 200 displays the first object using the third portion 210 c of the first microdisplay 204 a and the third portion 212 c of the second microdisplay 204 b .
- the device 200 displays the second object using the first portion 210 a of the first microdisplay 204 a and the first portion 210 a of the second microdisplay 204 b .
- a user may view the first microdisplay 204 a using their left eye and view the second microdisplay 204 b using their right eye.
- the device 200 displaying the first object using portions of the microdisplays (e.g., third portion 210 c of the first microdisplay 204 a and third portion 212 c of the second microdisplay 204 b ) on different planes than portions of microdisplays (e.g., first portion 210 a of the first microdisplay 204 a and first portion 212 a of the second microdisplay 204 b ) displaying the second object provides depth and reduces or eliminates the vergence-accommodation conflict.
- portions of the microdisplays e.g., third portion 210 c of the first microdisplay 204 a and third portion 212 c of the second microdisplay 204 b
- portions of microdisplays e.g., first portion 210 a of the first microdisplay 204 a and first portion 212 a of the second microdisplay 204 b
- the device 200 alters the display of the piece of 3D content based on the position and/or orientation of one or more portions of a microdisplay. For example, the device 200 may alter the 3D content displayed on the second portion 210 b of the first microdisplay 204 a and the second portion 212 b of the second microdisplay 204 b based on the orientations of the second portion 210 b of the first microdisplay 204 a and the second portion 212 b of the second microdisplay 204 b.
- one or more portions of a microdisplay corresponds to a fiber optic fabric.
- the second portion 210 b of the first microdisplay 204 a and the second portion 212 b of the second microdisplay 204 b may correspond to fiber optic fabric.
- the fiber optic fabric is connected between microdisplays and/or portions of microdisplays.
- a fiber optic fabric e.g., the second portion 210 b of the first microdisplay 204 a
- fiber optic fabric ensures that there are no observable gaps between microdisplays.
- boundaries of the microdisplays and/or portions of microdisplays are propagated to the fiber optic fabric.
- fiber optic fabric e.g., the second portion 210 b of the first microdisplay 204 a
- fiber optic fabric may be connected to a first edge of the first portion 210 a of the first microdisplay 204 a and a first edge of the third portion 210 c of the first microdisplay 204 a .
- the first edge of the first portion 210 a of the first microdisplay 204 a may display one or more colors.
- the one or more colors may be propagated onto the fiber optic fabric (e.g., the second portion 210 b of the first microdisplay 204 a ).
- the fiber optic fabric e.g., the second portion 210 b of the first microdisplay 204 a ) displaying the one or more colors of the first edge of the first portion 210 a of the first microdisplay 204 a provides continuity between the first portion 210 a of the first microdisplay 204 a and the third portion 210 c of the first microdisplay 204 b.
- FIGS. 3 A- 3 C show illustrative diagrams of a device 300 comprising a lens 302 .
- the device 300 is a component of a second device (e.g., device 100 ).
- a user 304 may use their left eye to look through the lens 302 of the device 300 and may use their right eye to look through a different lens (e.g., second lens 102 b ) of a different component of the second device (e.g., device 100 ).
- a different lens e.g., second lens 102 b
- additional components may use the same or similar methodologies and technology described herein.
- the device 300 is a component of a head-mounted device used to display 3D content.
- the device 300 may also comprise a first microdisplay 306 a , a second microdisplay 306 b , a third microdisplay 306 c , a fourth microdisplay 306 d , and a fifth microdisplay 306 e .
- the user 304 views one or more microdisplays by looking through the lens 302 .
- the lens 302 provides a field of view 310 that is visible to the user 304 .
- the device 300 may also comprise one or more members (not shown) used to connect the microdisplays to a housing 308 . Although only five microdisplays are shown, any number of microdisplays may be housed within the device 300 . In some embodiments, not all components of the device 300 are shown to avoid overcomplicating the drawing.
- the device 300 displays a piece of 3D content on the microdisplays.
- the device 300 may determine that the piece of 3D content comprises a scene with a plurality of objects using one or more of the methodologies described herein.
- the piece of 3D content may comprise a first object 312 a at a first location within the piece of 3D content, a second object 312 b at a second location within the piece of 3D content, a third object 312 c at a third location within the piece of 3D content, a fourth object 312 d at a fourth location within the piece of 3D content, and a fifth object 312 e at a fifth location within the piece of 3D content.
- the device 300 may change the positions of one or more microdisplays according to the locations of the plurality of objects within the scene. For example, the device 300 may determine that the first microdisplay 306 a will display the first object 312 a and change the position of the first microdisplay 306 a to a first position based on the location of the first object 312 a within the piece of 3D content. The device 300 may also determine that the second microdisplay 306 b will display the second object 312 b and change the position of the second microdisplay 306 b to a second position based on the location of the second object 312 b within the piece of 3D content.
- the device 300 may also determine that the third microdisplay 306 c will display the third object 312 c and change the position of the third microdisplay 306 c to a third position based on the location of the third object 312 c within the piece of 3D content.
- the device 300 may also determine that the fourth microdisplay 306 d will display the fourth object 312 d and change the position of the fourth microdisplay 306 d to a fourth position based on the location of the fourth object 312 d within the piece of 3D content.
- the device 300 may also determine that the fifth microdisplay 306 e will display the fifth object 312 e and change the position of the fifth microdisplay 306 e to a fifth position based on the location of the fifth object 312 e within the piece of 3D content.
- the device 300 may receive an input during the display of the 3D content.
- the input may correspond to the device 300 changing positions and/or orientations.
- the device 300 may receive an input when the user 304 moves their head (e.g., pan, tilt, roll, walk in a direction, etc.).
- the device 300 may comprise one or more sensors (e.g., accelerometers, gyroscopes, image sensors, depth sensors, light detection and ranging (LiDAR) sensors, and/or any similar such sensor) able to detect a change in position and/or orientation of the device 300 .
- the one or more sensors may be external to the device 300 .
- the device 300 may receive a first input from the one or more sensors.
- the input may correspond to a change in the gaze of the user 304 .
- the device 300 may use one or more sensors to perform eye-tracking on the user 304 to determine that the user 304 is looking to the right.
- the user 304 looking to the right causes the device 304 to move one or more microdisplays and/or change what one or more microdisplays are displaying.
- the device 300 updates the position of one or more microdisplays based on a received input. For example, the device 300 may display the piece of 3D content using all the microdisplays 306 a - 306 e . In some embodiments, the device 300 controls or positions the one or more microdisplays to display a part of the piece of 3D content that are outside a present field of view 310 of the user 304 .
- the device 300 may control a position of the first microdisplay 306 a to display the first object 312 a when the field of view 310 shifts to include the first object 312 a , and may control a position of the fifth microdisplay 306 e to display the fifth object 312 e when the field of view 310 shifts to include the fifth object 312 e .
- the device 300 may receive a first input from a first sensor (e.g., accelerometer) indicating that the device 300 has changed positions. In response to the first input, the device 300 may move the positions of one or more microdisplays according to the change in position of the device 300 .
- a first sensor e.g., accelerometer
- the device 300 may change the microdisplays from the starting positions (e.g., shown in FIG. 3 A ) to updated positions (e.g., shown in FIG. 3 B ) in response to the first input.
- the device 300 changing the positions of the microdisplays in response to the first input simulates a 3D environment for the user 304 .
- the user 304 was unable to view the fifth object 312 e because the fifth object 312 e was outside the field of view 310 of the user 304 .
- the device 300 When the user 304 pans their head to the right, the device 300 then changes the positions of the microdisplays so that a new object (e.g., fifth object 312 e displayed by the fifth microdisplay 306 e ) enters the field of view 310 of the user 304 simulating a 3D environment.
- a new object e.g., fifth object 312 e displayed by the fifth microdisplay 306 e
- the received input may include changes in multiple directions.
- the user 304 may pan their head to the right and walk forward or otherwise provide an input to move their field of view forward (e.g., toward the displayed objects) at the same time.
- the device 300 may receive a second input from a sensor (e.g., accelerometer) or other controller indicating that the device 300 has changed positions and orientations.
- the device 300 may move the positions of one or more microdisplays according to the change in position and orientation of the device 300 .
- the device 300 may change the microdisplays from the starting positions (e.g., shown in FIG. 3 A ) to second updated positions (e.g., shown in FIG.
- FIG. 3 C displays the fifth microdisplay 306 e entering the field of view 310 of the user 304 , as well as the microdisplays being closer to the lens 302 , which corresponds to a depth of the objects or depth of a virtual plane of the objects being reduced.
- the second updated positions result in the user 304 being able to see the fifth object 312 e .
- the second updated positions also result in the user 304 observing the third object 312 c , fourth object 312 d , and fifth object 312 e moving closure. Accordingly, the second updated positions simulate the user 304 panning their head to the right while walking or moving toward the displayed objects.
- the device 300 uses the received input to determine a speed for changing the position of one or more of the microdisplays. For example, if the received input indicates that the user 304 pans their head quickly to the right, the device 300 may change the microdisplays from the starting positions to the updated positions at a first speed, and if the received input indicates that the user 304 pans their head slowly to the right, the device 300 may change the microdisplays from the starting positions to the updated positions at a second speed that is slower than the first speed.
- the received input comprises acceleration and/or velocity measurements related to the device 300 .
- the received input comprises position information (e.g., coordinates of the device 300 ) related to the device 300 .
- the device 300 is programed to change the positions of microdisplays at different planes at different speeds. For example, in response to receiving an input, the device 300 may change the third microdisplay 306 c from the starting position (e.g., as shown in FIG. 3 A ) to the updated position (e.g., as shown in FIG. 3 B ) at a first speed and may change the fifth micro display 306 e from the starting position (e.g., as shown in FIG. 3 A ) to the updated position (e.g., as shown in FIG. 3 B ) at a second faster speed. In some embodiments, the device 300 is programed to change the positions of the microdisplays at different speeds to provide a real-world experience.
- the device 300 may change the position of the fifth microdisplay 306 e at a first speed so that the user 304 observes a first object, displayed on the fifth microdisplay 306 e , moving as if the user 304 was viewing the object in the real world.
- the device 300 updates the objects being displayed by the microdisplays in response to an input.
- the microdisplays may display the objects shown in FIG. 3 A .
- the device 300 may receive a third input (e.g., the user 304 stepping or moving to the right).
- the device 300 may change the display of the microdisplays so that the second microdisplay 306 b displays the third object 312 c , the third microdisplay 306 c displays the fourth object 312 d , and the fourth microdisplay 306 d displays the fifth object 312 e .
- the device 300 may also update the positions of one or more microdisplays based on the objects displayed.
- the device 300 may change the position of the second microdisplay 306 b based on the location of the third object 312 c within the piece of 3D content.
- the device 300 may also change the position of the third microdisplay 306 c based on the location of the fourth object 312 d within the piece of 3D content.
- the device 300 may also change the position of the fourth microdisplay 306 d based on the location of the fifth object 312 e within the piece of 3D content.
- the device 300 updates the positions of one or more microdisplays and/or the objects being displayed by one or more microdisplays in response to an input.
- the microdisplays may display the objects shown in FIG. 3 A .
- the device 300 may determine that the scene depicted in the piece of 3D content comprises one or more objects behind the user 304 .
- the scene depicted in the piece of 3D content may comprise a sixth object, a seventh object, and an eighth object located behind the user 304 .
- the device 300 may receive a fourth input (e.g., the user 304 turns around).
- the device 300 may change the display and/or positions of the microdisplays to reflect the objects (e.g., sixth object, seventh object, and eighth object) that have entered into the field of view 310 of the user 304 .
- the device 300 may change the display of the microdisplays so that the second microdisplay 306 b displays the sixth object, the third microdisplay 306 c displays the seventh object, and the fourth microdisplay 306 d displays the eighth object.
- the device 300 may also change the position of the second microdisplay 306 b to a first updated position based on the location of the sixth object within the piece of 3D content.
- the device 300 may also change the position of the third microdisplay 306 c to a second updated position based on the location of the seventh object within the piece of 3D content.
- the device 300 may also change the position of the fourth microdisplay 306 d to a third updated position based on the location of the eighth object within the piece of 3D content.
- the device 300 may change the positions of the microdisplays using one or more members.
- the first microdisplay 306 a may be attached to the housing 308 using a first member.
- the device 300 may us the first member to change the starting position of first microdisplay 306 a (e.g., as shown in FIG. 3 A ) to an updated position (e.g., as shown in FIG. 3 B ).
- the one or members are telescopic support rods, micropillars, fluid chambers, and/or similar such members.
- the device 300 uses a combination of gears and sliders to change the positions of one or more microdisplays.
- the device 300 may also respond to other movements. For example, the device 300 may receive a second input indicating that the user 304 tilted their head down. The device 300 may move the positions of the microdisplays vertically in the y-direction at a first speed according to the received second input. In some embodiments, the device 300 may comprise additional microdisplays (not shown) below and/or above the displayed microdisplays. In such an embodiment, the additional microdisplays may be outside the field of view 310 of the user 304 in the y-direction. If the devices 300 moves the positions of the microdisplays and the additional microdisplays vertically in the y-direction, in response to the second input, then one or more of the additional microdisplays may become visible to the user 304 .
- FIG. 4 shows an illustrative diagram of parameters used to determine positions for one or more portions of a display of a device 400 , in accordance with embodiments of the disclosure.
- the device 400 comprises a lens 402 .
- the device 400 is a component of a second device (e.g., device 100 ).
- a user 404 may use their left eye to look through the lens 402 of the device 400 and may use their right eye to look through a different lens (e.g., second lens 102 b ) of a different component of the second device (e.g., device 100 ).
- a different lens e.g., second lens 102 b
- additional components may use the same or similar methodologies and technology described herein.
- the device 400 is a component of a head-mounted device used to display 3D content.
- the device 400 may comprise a first microdisplay 406 a , a second microdisplay 406 b , and a third microdisplay 406 c .
- the device 400 may also comprise one or more members (not shown) used to connect the microdisplays to a housing 408 . Although only three microdisplays are shown, any number of microdisplays may be housed within the device 400 . In some embodiments, not all components of the device 400 are shown to avoid overcomplicating the drawing.
- the device 400 receives a piece of 3D content.
- the piece of the 3D content may depict a scene comprising a plurality of objects from a first viewpoint.
- a piece of 3D content may depict a scene of a landscape from the viewpoint of the camera filming the landscape.
- the scene may comprise a first object 410 a , a second object 410 b , and a third object 410 c .
- the plurality of objects are located at different positions within the scene.
- the first object 410 a may be a first distance 412 a from the first viewpoint
- the second object 410 b may be a second distance 412 b from the first viewpoint
- the third object 410 c may be a third distance 412 c from the first viewpoint.
- the device 400 may use the focal length of the lens 402 and the positions of the plurality of objects within the scene to determine what locations the one or more microdisplays should be changed to when displaying the piece of 3D content.
- f focal length of the lens 402 .
- d distance between lens 402 and a microdisplay.
- z distance between the viewpoint and an object in the piece of 3D content.
- the device 400 uses Equation (1) to determine a location for one or more microdisplays. For example:
- d 1 ( f * z 1 ) / ( z 1 - f ) ( 2 )
- f focal length of the lens 402
- d 1 distance 414 a between lens 402 and the first microdisplay 406 a
- z 1 distance 412 a between the viewpoint and the first object 410 a in the piece of 3D content.
- d 2 ( f * z 2 ) / ( z 2 - f ) ( 3 )
- f focal length of the lens 402 .
- d 2 distance 414 b between lens 402 and the second microdisplay 406 b.
- z 2 distance 412 b between the viewpoint and the second object 410 b in the piece of 3D content.
- d 3 ( f * z 3 ) / ( z 3 - f ) ( 4 )
- f focal length of the lens 402 .
- d 3 distance 414 c between lens 402 and the third microdisplay 406 c.
- z 3 distance 412 c between the viewpoint and the third object 410 c in the piece of 3D content.
- the lens 402 has a focal length of 30 millimeters (mm) and the first distance 412 a between the viewpoint and the first object 410 a in the piece of 3D content is 100 mm.
- the device may determine that the distance 414 a between the first microdisplay 406 a and the lens 402 should be 42.85 mm.
- the device 400 may cause the first microdisplay 406 a to change from a starting position to the calculated distance 414 a to display the first object 410 a for the user 404 .
- the first object 410 a may change distances from the first viewpoint during the piece of 3D content.
- the piece of 3D content may comprise a plurality of segments.
- the plurality of segments represent various portions and/or parts of the piece of 3D content.
- a first segment may relate to the beginning of the piece of 3D content and a second segment may relate to the end of the piece of 3D content.
- a first segment may relate to a first part of the piece of 3D content displaying virtual objects in a first configuration and a second segment may relate to a second part of the piece of 3D content displaying virtual objects in a second configuration.
- each segment corresponds to a length of time of the piece of 3D content.
- the first object 410 a may be a first distance 412 a from the first viewpoint during a first segment of the piece of 3D content and may be a different distance from the first viewpoint during a second segment of the piece of 3D content.
- the device 400 uses the focal length of the lens 402 and the different distance from the first viewpoint during the second segment of the piece of 3D content to determine an updated location for the first microdisplay 406 a .
- the device 400 may use Equation (2) to determine an updated distance between the first microdisplay 406 a and the lens 402 to be 41.25 mm.
- the device 400 may cause the first microdisplay 406 a to change from the first distance 414 a to the updated distance to display the first object 410 a for the user 404 during the second segment of the piece of the 3D content.
- the device 400 determines a speed at which to move the one or more microdisplays. For example, the device 400 may determine a speed at which to move the first microdisplay 406 a from the first distance 414 a to the updated distance to display the first object 410 a for the user 404 during the second segment of the piece of the 3D content. The device may determine a first speed corresponding to the change in which the first object 410 a changes from the first distance 412 a from the first viewpoint to the updated distance.
- Equation (2) can be modified so that:
- d 1 ′ ( f ⁇ ( z 1 + S ) ) / ( ( z 1 + S ) - f ) ( 5 )
- d 1 ′ new distance between lens 402 and the first microdisplay 406 a.
- z 1 distance 412 a between the viewpoint and the first object 410 a in the piece of 3D content.
- S distance per second in which the first object 410 a changes from the first distance 412 a from the first viewpoint to the updated distance from the first viewpoint.
- Equation (5) and Equation (2) may be used to find a second speed corresponding to the change of the first microdisplay 406 a from the distance 414 a between the first microdisplay 406 a and the lens 402 to the new distance between the first microdisplay 406 a and the lens 402 .
- a second speed corresponding to the change of the first microdisplay 406 a from the distance 414 a between the first microdisplay 406 a and the lens 402 to the new distance between the first microdisplay 406 a and the lens 402 .
- Equation (6) can be simplified to:
- the device 400 changes the focal length of the lens 402 to reduce the speed at which the device 400 would have to move the first microdisplay 406 a to display the first object 410 a changing distances relative to the first viewpoint of the piece of 3D content. In some embodiments, the device 400 manipulates the display of the piece of 3D content so that the distance 412 a between the viewpoint and the first object 410 a in the piece of 3D content is larger.
- the larger distance between the viewpoint and the first object 410 a reduces the speed at which the first microdisplay 406 a would have to move the first microdisplay 406 a to display the first object 410 a changing distances relative to the first viewpoint of the piece of 3D content.
- the device uses a combination of methodologies to reduce the speed at which the first microdisplay 406 a would have to move to display the first object 410 a changing distances relative to the first viewpoint of the piece of 3D content.
- FIGS. 5 A and 5 B show illustrative diagrams of displays comprising multiple portions, in accordance with embodiments of the disclosure.
- FIG. 5 A shows a first display 500
- FIG. 5 B shows a second display 502 .
- the displays are used to display 3D content.
- the first display 500 may be part of a head-mounted device used for displaying 3D content.
- the second display 502 may be a display used by a television, laptop, smartphone, and/or similar such device to display 3D content.
- the displays comprise a plurality of microdisplays.
- the first display 500 may comprise 12 microdisplays and the second display 502 may comprise over 200 microdisplays.
- the displays comprise portions that are not connected to each other.
- the first display 500 may comprise 12 microdisplays, where each microdisplay may change positions independent of each other (e.g., microdisplays 104 a - 104 d shown in FIG. 1 A ).
- the displays comprise portions that are not connected to each other.
- the second display 502 may be one continuous flexible display.
- the displays are associated with a single lens.
- the first display 500 may only be visible through a single lens (e.g., first lens 102 a shown in FIG. 1 A).
- the displays are visible from two lenses.
- the second display 502 may be visible through two lenses (e.g., first lens 102 a and second lens 102 b shown in FIG. 1 A ).
- the displays and/or the portions of the displays have rectangular cross-sections. Although rectangular cross-sections are shown, other shapes of displays and/or portions of displays may be used. For example, the displays and/or the portions of the displays may have circular cross-sections, triangular cross-sections, hexagonal cross-sections, and/or similar such cross-sections. In some embodiments, the displays and/or the portions of the displays have the same or similar size and/or shaped cross-sections. In some embodiments, the shape and/or size of the cross-sections of the displays and/or the portions of the displays vary.
- the first display 500 may comprise four microdisplays with a triangular cross-section, four microdisplays with a circular cross-section, four microdisplays with a rectangular cross-section of a first size, and a four microdisplays with a rectangular cross-section of a second size.
- the displays are attached to a housing using one or more members.
- the first display 500 may be attached to a housing using one or more telescopic support rods.
- FIGS. 6 A and 6 B show illustrative diagrams of a system 600 for manipulating a display so that portions of the display are located at varying planes, in accordance with embodiments of the disclosure.
- a microdisplay 602 is coupled to a platform 606 . Although only one microdisplay is shown, any number of microdisplays may be coupled to the platform 606 and/or use a mechanism similar to what is displayed in FIGS. 6 A and 6 B .
- the microdisplay 602 is a portion of a microdisplay (e.g., third portion 210 c of the first microdisplay 204 a in FIG. 2 B ).
- FIG. 6 A displays the microdisplay 602 in a first position.
- FIG. 6 B displays the microdisplay 602 in a second position.
- the microdisplay 602 is coupled to a first member 604 .
- the first member 604 extends the microdisplay 602 away from the platform 606 .
- the first member 604 may extend in the direction shown in FIGS. 6 A and 6 B .
- the first member 604 retracts, bringing the microdisplay 602 toward the platform 606 .
- the system 600 utilizes a MEMS to change the microdisplay 602 from the first portion to the second position.
- the system 600 manipulates fluid within the first member 604 and/or the second member 608 to change the position of the microdisplay 602 .
- FIG. 6 B displays the microdisplay 602 in a second position.
- the microdisplay 602 is coupled to the first member 604 and the second member 608 .
- the second member 608 extends the microdisplay 602 away from the platform 606 .
- the first member 604 and/or second member 608 change the position of the microdisplay 602 based on a piece of 3D content. For example, a device may determine that a first object within the piece of 3D content is located on a first plane. The device may then determine one or more locations for the microdisplay 602 according to the plane associated with the first object.
- the device may calculate the second position for the microdisplay 602 to display the first object based on a focal length of a lens (e.g., first lens 102 a ) and the first object being located on the first plane within the piece of 3D content.
- the device may then use the first member 604 and the second member 608 to change the position of the microdisplay 602 from the first position (e.g., as shown FIG. 6 A ) to a second position (e.g., as shown in FIG. 6 B ).
- the microdisplay 602 at the second position displays the first object of the piece of 3D content.
- FIGS. 7 A- 7 D show other illustrative diagrams of a system 700 for manipulating a display so that portions of the display are located at varying planes, in accordance with embodiments of the disclosure.
- a microdisplay 702 is attached to a member 704 and the member 704 is attached to a platform 706 . Although only one microdisplay 702 and one member 704 are shown, any number of microdisplays may be coupled to the platform 706 using any number of members.
- the microdisplay 702 is a portion of a microdisplay (e.g., the third portion 210 c of the first microdisplay 204 a in FIG. 2 B ).
- FIG. 7 A displays the microdisplay 702 in a first position.
- FIG. 7 B displays the microdisplay 702 in a second position.
- the member 704 extends the microdisplay 702 away from the platform 706 .
- the member 704 may extend in the direction shown in FIGS. 7 A and 7 B .
- the member 704 retracts, bringing the microdisplay 702 toward the platform 706 .
- the system 700 utilizes a MEMS to change the microdisplay 702 from the first position to the second position.
- the member 704 may be a telescopic support rod.
- the system 700 manipulates fluid within the member 704 to change the position of the microdisplay 702 .
- the system 700 uses a portion 708 of the member 704 to change the position of the microdisplay 702 .
- the portion 708 of the member 704 may be used to tilt, rotate, extend, and/or retract the microdisplay 702 in relation to the member 704 .
- FIG. 7 C displays the system 700 using the portion 708 of the member 704 to change the microdisplay 702 from the second position (shown in FIG. 7 B ) to a third position.
- the system 700 changes the microdisplay 702 to the third position by tilting and rotating the microdisplay 702 .
- tilting and/or rotating the microdisplay 702 using the portion 708 of the member 704 provides improved display of the piece of 3D content. For example, tilting and/or rotating the microdisplay 702 using the portion 708 of the member 704 may reduce or eliminate holes between two or more microdisplays.
- FIG. 7 D displays the microdisplay 702 in a fourth position.
- the system 700 comprises a first track 710 and a second track 712 . Although only two tracks are shown, any number of tracks and/or any shape of tracks may be user. For example, there may be one or more circular tracks, one or more rectangular tracks, etc.
- the system 700 slides the member 704 in a first direction along the first track 710 to change the microdisplay 702 from the second position (shown in FIG. 7 B ) to the fourth position (shown in FIG. 7 D ). In some embodiments, sliding the microdisplay 702 along one or more tracks improves the display of the piece 3D content.
- a device e.g., device 100 , device 200 , user equipment device 1100 , etc. uses one or more of the techniques and/or technologies described herein to change the position of a portion of a microdisplay.
- FIGS. 8 A and 8 B shows illustrative diagrams of a lens 802 used with a device for manipulating a stereoscopic display so that portions of the stereoscopic display are located at varying planes, in accordance with embodiments of the disclosure.
- a lens 802 is coupled to a platform 804 .
- any number of lenses may be coupled to the platform 804 and/or use the mechanism displayed in FIGS. 8 A and 8 B .
- a first lens e.g., first lens 102 a shown in FIG. 1 A
- a second lens e.g., second lens 102 b shown in FIG. 1 A
- FIG. 8 A displays the lens 802 in a first shape.
- FIG. 8 B displays the lens 802 in a second shape.
- a device changes the shape of the lens 802 using a MEMS component, a micro-piezoelectric component, and/or other electro-mechanical controllers.
- the lens 802 may be coupled to one or more members.
- the one or more members may expand or contract the lens 802 in one or more directions.
- the device may change the position of the lens 802 .
- the device may extend the lens 802 away from the platform 804 using one or more members.
- the device may tilt the lens 802 .
- the device may change the positions of the lens 802 from a first depth within the platform 804 to a second depth within the platform 804 .
- FIGS. 9 A and 9 B show other illustrative diagrams of a lens 902 used with a device for manipulating a stereoscopic display so that portions of the stereoscopic display are located at varying planes, in accordance with embodiments of the disclosure.
- the lens 902 is a liquid lens and has a top surface 906 .
- the top surface 906 is liquid or a film.
- the lens 902 may be coupled to a platform 904 . Although one lens is shown, any number of lenses may be coupled to the platform 904 and/or use the mechanisms displayed in FIGS. 9 A and 9 B .
- a first lens e.g., first lens 102 a shown in FIG. 1 A
- a second lens e.g., second lens 102 b shown in FIG. 1 A
- FIG. 9 A displays the lens 902 in a first shape.
- FIG. 9 B displays the lens 904 in a second shape.
- a device changes the lens 902 from the first shape to the second shape by applying a force 908 to the lens 902 .
- the device may constrict the lens 902 so that the top surface 906 becomes more convex.
- the device changes the shape of the lens 902 using electrowetting, shape-changing polymers, acusto-optical tuning, and/or similar such methodologies.
- the device changes the focal length of the lens 902 to reduce the speed at which the device would have to move one or more microdisplays to display an object changing distances relative to a first viewpoint of a piece of 3D content. For example, if the speed at which the device would have to move one or more microdisplays is above a threshold speed, the device may change the lens 902 from the first shape (e.g., as shown in FIG. 9 A ) to the second shape (e.g., as shown in FIG. 9 A ).
- the device may change the lens 902 from a first position (e.g., first depth within the platform 904 ) to a second position (e.g., second depth within the platform 904 ).
- FIGS. 10 - 11 describe exemplary devices, systems, servers, and related hardware for manipulating a display (e.g., stereoscopic display) so that portions of the display are located at varying planes, in accordance with some embodiments.
- a display e.g., stereoscopic display
- a user may utilize more than one type of user equipment device 1002 and more than one of each type of user equipment device.
- the user equipment device 1002 may be an XR device (e.g., head-mounted display) and/or an audio output device.
- the user equipment device 1002 may be a television, smartphone, laptop, desktop, tablet, and/or any other type of computer able to display content.
- there may be paths between user equipment devices so that the devices may communicate directly with each other via communications paths, as well as other short-range point-to-point communications paths, such as USB cables, IEEE 1394 cables, wireless paths (e.g., Bluetooth, infrared, IEEE 802-11x, etc.), or other short-range communication via wired or wireless paths.
- the user equipment devices may also communicate with each other directly through an indirect path via the communications network 1006 .
- the user equipment devices may be coupled to communications network 1006 .
- the user equipment device 1002 is coupled to the communications network 1006 via communications path 1004 .
- a media content source 1012 is coupled to the communications network 1006 via communications path 1008 and a server 1014 is coupled to the communications network 1006 via communication path 1010 .
- the communications network 1006 may be one or more networks including the Internet, a mobile phone network, mobile voice or data network (e.g., a 4G, 5G, or LTE network), cable network, public switched telephone network, or other types of communications network or combinations of communications networks.
- the paths may separately or in together with other paths include one or more communications paths, such as, a satellite path, a fiber-optic path, a cable path, a path that supports Internet communications (e.g., IPTV), free-space connections (e.g., for broadcast or other wireless signals), or any other suitable wired or wireless communications path or combination of such paths.
- one or more paths can be a wireless path. Communication with the user equipment device may be provided by one or more communications paths but is shown as a single path in FIG. 10 to avoid overcomplicating the drawing.
- the system 1000 also includes media content source 1012 , and server 1014 , which can be coupled to any number of databases providing information to the user equipment devices.
- media content source 1012 and server 1014 may have access to stereoscopic data, augmentation data, 2D and/or 3D mapping data, and/or similar such information.
- the media content source 1012 represents any computer-accessible source of content, such as a storage for audio content, metadata, or, similar such information.
- the server 1014 may store and execute various software modules for manipulating a display so that portions of the display are located at varying planes functionality.
- the user equipment device 1002 , media content source 1012 , and server 1014 may store metadata associated with media content.
- media content, content item, piece of media, piece of 3D content may be used interchangeably.
- FIG. 11 shows a generalized embodiment of a user equipment device 1100 , in accordance with one embodiment.
- the user equipment device 1100 is an example of the user equipment devices described in FIGS. 1 - 10 (e.g., device 102 ).
- the user equipment device 1100 may receive content and data via input/output (I/O) path 1102 .
- the I/O path 1102 may provide audio content (e.g., broadcast programming, on-demand programming, Internet content, content available over a local area network (LAN) or wide area network (WAN), and/or other content) and data to control circuitry 1104 , which includes processing circuitry 1106 and a storage 1108 .
- audio content e.g., broadcast programming, on-demand programming, Internet content, content available over a local area network (LAN) or wide area network (WAN), and/or other content
- control circuitry 1104 which includes processing circuitry 1106 and a storage 1108 .
- the control circuitry 1104 may be used to send and receive commands, requests, and other suitable data using the I/O path 1102 .
- the I/O path 1102 may connect the control circuitry 1104 (and specifically the processing circuitry 1106 ) to one or more communications paths. I/O functions may be provided by one or more of these communications paths but are shown as a single path in FIG. 11 to avoid overcomplicating the drawing.
- the control circuitry 1104 may be based on any suitable processing circuitry such as the processing circuitry 1106 .
- processing circuitry should be understood to mean circuitry based on one or more microprocessors, microcontrollers, digital signal processors, programmable logic devices, field-programmable gate arrays (“FPGAs”), application-specific integrated circuits (“ASICs”), etc., and may include a multi-core processor (e.g., dual-core, quad-core, hexa-core, or any suitable number of cores) or supercomputer.
- processing circuitry may be distributed across multiple separate processors or processing units, for example, multiple of the same type of processing units (e.g., two Intel Core i7 processors) or multiple different processors (e.g., an Intel Core i5 processor and an Intel Core i7 processor).
- the manipulating a display so that portions of the display are located at varying planes functionality can be at least partially implemented using the control circuitry 1104 .
- the manipulating a display so that portions of the display are located at varying planes functionality described herein may be implemented in or supported by any suitable software, hardware, or combination thereof.
- the providing of stereoscopic data, augmentation data, 2D data, and/or 3D data can be implemented on user equipment, on remote servers, or across both.
- control circuitry 1104 may include communications circuitry suitable for communicating with one or more servers that may at least implement the described manipulating of a display so that portions of the display are located at varying planes functionality.
- the instructions for carrying out the above-mentioned functionality may be stored on the one or more servers.
- Communications circuitry may include a cable modem, an integrated service digital network (“ISDN”) modem, a digital subscriber line (“DSL”) modem, a telephone modem, Ethernet card, or a wireless modem for communications with other equipment, or any other suitable communications circuitry. Such communications may involve the Internet or any other suitable communications networks or paths.
- communications circuitry may include circuitry that enables peer-to-peer communication of user equipment devices, or communication of user equipment devices in locations remote from each other (described in more detail below).
- Memory may be an electronic storage device provided as the storage 1108 that is part of the control circuitry 1104 .
- the phrase “electronic storage device” or “storage device” should be understood to mean any device for storing electronic data, computer software, or firmware, such as random-access memory, read-only memory, hard drives, optical drives, digital video disc (“DVD”) recorders, compact disc (“CD”) recorders, BLU-RAY disc (“BD”) recorders, BLU-RAY 3D disc recorders, digital video recorders (“DVR”, sometimes called a personal video recorder, or ���PVR”), solid-state devices, quantum storage devices, gaming consoles, gaming media, or any other suitable fixed or removable storage devices, and/or any combination of the same.
- DVD digital video disc
- CD compact disc
- DVR digital video recorders
- PVR personal video recorder
- the storage 1108 may be used to store various types of content described herein. Nonvolatile memory may also be used (e.g., to launch a boot-up routine and other instructions). Cloud-based storage, described in relation to FIG. 10 , may be used to supplement the storage 1108 or instead of the storage 1108 .
- the control circuitry 1104 may include audio generating circuitry and tuning circuitry, such as one or more analog tuners, audio generation circuitry, filters or any other suitable tuning or audio circuits or combinations of such circuits.
- the control circuitry 1104 may also include scaler circuitry for upconverting and down converting content into the preferred output format of the user equipment device 1100 .
- the control circuitry 1104 may also include digital-to-analog converter circuitry and analog-to-digital converter circuitry for converting between digital and analog signals.
- the tuning and encoding circuitry may be used by the user equipment device 1100 to receive and to display, to play, or to record content.
- the circuitry described herein including, for example, the tuning, audio generating, encoding, decoding, encrypting, decrypting, scaler, and analog/digital circuitry, may be implemented using software running on one or more general purpose or specialized processors. If the storage 1108 is provided as a separate device from the user equipment device 1100 , the tuning and encoding circuitry (including multiple tuners) may be associated with the storage 1108 .
- the user may utter instructions to the control circuitry 1104 , which are received by the microphone 1116 .
- the microphone 1116 may be any microphone (or microphones) capable of detecting human speech.
- the microphone 1116 is connected to the processing circuitry 1106 to transmit detected voice commands and other speech thereto for processing.
- voice assistants e.g., Siri, Alexa, Google Home and similar such voice assistants
- the user equipment device 1100 may optionally include an interface 1110 .
- the interface 1110 may be any suitable user interface, such as a remote control, mouse, trackball, keypad, keyboard, touch screen, touchpad, stylus input, joystick, or other user input interfaces.
- a display 1112 may be provided as a stand-alone device or integrated with other elements of the user equipment device 1100 .
- the display 1112 may be a touchscreen or touch-sensitive display.
- the interface 1110 may be integrated with or combined with the microphone 1116 .
- the interface 1110 When the interface 1110 is configured with a screen, such a screen may be one or more of a monitor, a television, a liquid crystal display (“LCD”), active matrix display, cathode ray tube display, light-emitting diode display, organic light-emitting diode display, quantum dot display, or any other suitable equipment for displaying visual images.
- the interface 1110 may be HDTV-capable.
- the display 1112 may have one or more portions that can be adjusted to varying planes.
- the control circuitry 1104 may utilize a MEMS to adjust a first portion of the display 1112 to a first plane.
- a speaker 1114 may be controlled by the control circuitry 1104 .
- the speaker (or speakers) 1114 may be provided as integrated with other elements of user equipment device 1100 or may be a stand-alone unit.
- the display 1112 may be outputted through speaker 1114 .
- the user equipment device 1100 of FIG. 11 can be implemented in system 1000 of FIG. 10 as user equipment device 1002 , but any other type of user equipment suitable for manipulating a display so that portions of the display are located at varying planes may be used.
- user equipment devices such as television equipment, computer equipment, wireless user communication devices, or similar such devices may be used.
- User equipment devices may be part of a network of devices.
- FIG. 12 is an illustrative flowchart of a process 1200 for manipulating a display so that portions of the display are located at varying planes in accordance with embodiments of the disclosure.
- Process 1200 may be executed by control circuitry 1104 on a user equipment device 1100 .
- control circuitry 1104 may be part of a remote server separated from the user equipment device 1100 by way of a communications network or distributed over a combination of both.
- the control circuitry is at a local server and/or within a head-mounted display device.
- instructions for executing process 1200 may be encoded onto a non-transitory storage medium (e.g., the storage 1108 ) as a set of instructions to be decoded and executed by processing circuitry (e.g., the processing circuitry 1106 ).
- Processing circuitry may, in turn, provide instructions to other sub-circuits contained within control circuitry 1104 , such as the encoding, decoding, encrypting, decrypting, scaling, analog/digital conversion circuitry, and the like.
- control circuitry 1104 such as the encoding, decoding, encrypting, decrypting, scaling, analog/digital conversion circuitry, and the like.
- the process 1200 or any step thereof, could be performed on, or provided by, any of the devices shown in FIGS. 1 A- 11 .
- the process 1200 and any of the following processes, are illustrated and described as a sequence of steps, it is contemplated that various embodiments of processes may be performed in any order or combination and need not include all the illustrated steps.
- control circuitry receives a piece of 3D content comprising a plurality of segments.
- the piece of 3D content is an image or plurality of images corresponding to a movie, television show, video game, and/or any other type of 3D content.
- the control circuitry receives the piece of 3D content from a server.
- control circuitry determines that a first segment of the plurality of segments comprises a 3D object, wherein the 3D object corresponds to a first location within the first segment.
- the first segment of the piece of 3D content may depict a landscape, where the 3D object (e.g., cloud) is located at first location.
- the control circuitry determines that the first segment of the piece of 3D content comprises the 3D object using metadata associated with the piece of 3D content.
- the control circuitry may receive metadata associated with the piece of 3D content, wherein the metadata indicates that the 3D content comprises 3D objects.
- the control circuitry may receive the metadata when the control circuitry receives the piece of 3D content at step 1202 .
- the metadata may also provide information about the 3D objects displayed in the piece of 3D content.
- the metadata may indicate location information relating to the 3D objects displayed in the piece of 3D content.
- the location information may comprise coordinates related to the 3D objects displayed in the piece of 3D content, indicators specifying plane information related to the 3D objects displayed in the piece of 3D content, and/or similar such information.
- the piece of 3D content comprises the metadata.
- certain 3D objects receive a preference for determining the positions of one or more microdisplays.
- a segment may depict a person (e.g., first 3D object) standing in front of a tree (e.g., second 3D object).
- the control circuitry may determine that the first 3D object is more prominent in the segment than the second 3D object and assign the first 3D object a first ranking and the second 3D object a second ranking that is less than the first ranking.
- the control circuitry uses eye tracking to determine that the first 3D object is more prominent than the second 3D object.
- the control circuitry may determine that the first 3D object should have a higher ranking if the user's eyes spend more time looking at the first 3D object than the second 3D object.
- the control circuitry uses metadata to determine that the first 3D object is more prominent than the second 3D object.
- metadata associated with the piece of 3D content may indicate that the first 3D object has a higher ranking than the second 3D object.
- the control circuitry may determine that the first object and the second object are both going to be displayed by the same microdisplay (e.g., first microdisplay) but are located at different depths within the segment.
- the first segment may depict the first object as being 1 meter from the viewpoint of the segment and may depict the second object as being 20 meters from the viewpoint of the segment.
- the control circuitry may determine the position of the first microdisplay according to the location (e.g., 1 meter from the viewpoint of the segment) of the first object within the segment because the first object has a higher ranking than the second object.
- control circuitry changes a first microdisplay from a first position to a second position based on the first location of the 3D object within the first segment.
- the control circuitry calculates the second position for the first microdisplay using the focal length of a lens (e.g., the first lens 102 a shown in FIG. 1 A ) and the first location of the 3D object within the piece of 3D content.
- the control circuitry may use one or more equations (e.g., Equation 2) described above to determine a position for the first microdisplay relative to the lens.
- the control circuitry changes the position of the first microdisplay using one or more members.
- one or more members coupled to the first microdisplay may be telescopic support rods that can change the position of the first microdisplay.
- the one or more members may be fluid chambers that can change the position of the first microdisplay.
- the control circuitry utilizes a MEMS comprising one or more members to change the first microdisplay from the first position to the second position.
- control circuitry displays the first segment comprising the 3D object, wherein the first microdisplay is at the second position as the first segment is displayed.
- the control circuitry displays the first segment using a plurality of microdisplays.
- the second position of the first microdisplay is on a different plane relative to at least one microdisplay of the plurality of microdisplays.
- the first microdisplay may be a first vertical distance from a lens and a second microdisplay may be a second vertical distance from the lens.
- the control circuitry displaying the 3D object on the first microdisplay at the second position provides depth and reduces or eliminates the vergence-accommodation conflict.
- FIG. 13 is another illustrative flowchart of a process 1300 for manipulating a display so that portions of the display are located at varying planes in accordance with embodiments of the disclosure.
- control circuitry receives a piece of 3D content comprising a plurality of segments.
- the control circuitry uses the same or similar methodologies described above at step 1202 to receive the piece of 3D content comprising the plurality of segments.
- control circuitry determines that a first segment of the plurality of segments comprises a 3D object, wherein the 3D object corresponds to a first location within the first segment.
- the control circuitry uses the same or similar methodologies described above at step 1204 to determine that the first segment of the plurality of segments comprises the 3D object, wherein the 3D object corresponds to the first location within the first segment.
- control circuitry determines whether a microdisplay position change is required.
- the control circuitry identifies that one or more microdisplays of a plurality of microdisplays are associated with the 3D object. For example, the control circuitry may display the first segment of the piece of 3D content using 5 microdisplays, and a first microdisplay of the plurality of microdisplays may display the 3D object that is included in the first segment. Accordingly, the first microdisplay may be associated with the 3D object.
- control circuitry may display the first segment of the piece of 3D content using 5 microdisplays, a first microdisplay may display a first portion of the 3D object in the first segment, and a second microdisplay may display a second portion of the 3D object in the first segment. Accordingly, the first microdisplay and the second microdisplay may be associated with the 3D object.
- the control circuitry determines one or more positions for the one or more microdisplays associated with the 3D object to display the 3D object. For example, the control circuitry may determine that the first microdisplay is associated with the 3D object.
- the control circuitry may use a focal length of a lens (e.g., the first lens 102 a shown in FIG.
- control circuitry may determine that the first microdisplay and the second microdisplay are both associated with the 3D object.
- the control circuitry may use the focal length of the lens and the first location of the 3D object determined at step 1304 to determine the positions of the first microdisplay and the second microdisplay.
- the control circuitry compares the starting positions of the one or more microdisplays associated with the 3D content to corresponding calculated positions of the one or more microdisplay associated with the 3D content. For example, a starting position of the first microdisplay may be compared with the calculated position for the first microdisplay, and a starting position of the second microdisplay may be compared with the calculated position for the second microdisplay. If the control circuitry determines that one or more starting positions are different than the corresponding one or more calculated positions, then the control circuitry my determine that a microdisplay position change is required. If the control circuitry determines that a microdisplay position change is required, then the process 1300 continues to step 1308 . If the control circuitry determines that the microdisplay position change is not required, then the process 1300 continues to step 1310 .
- control circuitry changes a first microdisplay from a first position to a second position.
- the second position is the position calculated at step 1306 .
- the control circuitry may determine that the first microdisplay is associated with the 3D object.
- the control circuitry may use the focal length of a lens and the first location of the 3D object determined at step 1304 to determine the second position of the first microdisplay.
- the control circuitry changes the position of the first microdisplay using one or more members.
- one or more members coupled to the first microdisplay may be telescopic support rods that can change the position of the first microdisplay.
- the one or more members may be fluid chambers that can change the position of the first microdisplay.
- the control circuitry utilizes a MEMS component comprising one or more members to change the first microdisplay from the first position to the second position.
- control circuitry displays the first segment comprising the 3D object using a plurality of microdisplays.
- the second position of the first microdisplay is on a different plane relative to at least one microdisplay of the plurality of microdisplays.
- the first microdisplay may be a first vertical distance from a lens and a second microdisplay may be a second vertical distance from the lens.
- the control circuitry displaying the 3D object on the first microdisplay at the second position provides depth and reduces or eliminates the vergence-accommodation conflict.
- control circuitry determines that a second segment of the plurality of segments comprises the 3D object, wherein the 3D object corresponds to a second location within the second segment.
- the control circuitry uses the same or similar methodologies described at step 1204 when determining that the first segment comprises the 3D object to determine that the second segment comprises the 3D object.
- the control circuitry may receive metadata associated with the piece of 3D content, wherein the metadata indicates that the second segment of the 3D content comprises the 3D object.
- the metadata also provides information about the 3D objects displayed in the second segment of the piece of 3D content.
- the metadata may indicate location information relating to the 3D objects displayed in the second segment of the piece of 3D content.
- the location information may comprise coordinates related to the 3D objects displayed in the second segment of the piece of 3D content, indicators specifying plane information related to the 3D object displayed in the second segment of the piece of 3D content, and/or similar such information.
- the control circuitry uses the information about the 3D object displayed in the second segment to determine the second location.
- control circuitry determines whether a microdisplay position change is required.
- the control circuitry identifies that one or more microdisplays of a plurality of microdisplays are associated with the 3D object during the second segment.
- the control circuitry may display the second segment of the piece of 3D content using 5 microdisplays and the first microdisplay of the plurality of microdisplays may display the 3D object during the second segment. Accordingly, the first microdisplay may be associated with the 3D object for the second segment.
- the control circuitry may display the second segment of the piece of 3D content using 5 microdisplays and the first microdisplay may display the first portion of the 3D object and the second microdisplay may display the second portion of the 3D object. Accordingly, the first microdisplay and the second microdisplay may be associated with the 3D object during the second segment of the piece of 3D content.
- the control circuitry determines one or more positions for the one or more microdisplays associated with the 3D object during the second segment. For example, the control circuitry may determine that the first microdisplay is associated with the 3D object during the second segment. The control circuitry may use a focal length of the lens (e.g., the first lens 102 a shown in FIG. 1 A ) and the second location of the 3D object during the second segment, determined at step 1312 , to determine the position of the first microdisplay during the second segment of the piece of the 3D content. In another example, the control circuitry may determine that the first microdisplay and the second microdisplay are both associated with the 3D object during the second segment. The control circuitry may use the focal length of the lens and the second location of the 3D object during the second segment, determined at step 1304 , to determine the position of the first microdisplay and the second microdisplay.
- a focal length of the lens e.g., the first lens 102 a shown in FIG. 1 A
- the control circuitry may
- the control circuitry compares the positions of the one or more microdisplays associated with the 3D content during the previous segment (e.g., first segment) to corresponding calculated positions of the one or more microdisplay associated with the piece of 3D content during the second segment. For example, the position of the first microdisplay during the first segment may be compared with the calculated position for the first microdisplay during the second segment and the position of the second microdisplay during the first segment may be compared with the calculated position for the second microdisplay during the second segment. If the control circuitry determines that one or more positions of the microdisplays during the first segment are different than the corresponding one or more calculated positions during the second segment, then the control circuitry my determine that a microdisplay change is required. If the control circuitry determines that a microdisplay change is required, then the process 1300 continues to step 1316 . If the control circuitry determines that the microdisplay position change is not required, then the process 1300 continues to step 1318 .
- control circuitry changes the first microdisplay from the second position to a third position.
- the third position is the position calculated at step 1314 .
- the control circuitry may determine that the first microdisplay is associated with the 3D object for the second segment.
- the control circuitry may use the focal length of the lens and the second location of the 3D object during the second segment, determined at step 1314 , to determine the third position of the first microdisplay.
- the control circuitry changes the position of the first microdisplay using one or more members.
- one or more members coupled to the first microdisplay may be telescopic support rods that can change the position of the first microdisplay.
- the one or more members may be fluid chambers that can change the position of the first microdisplay.
- the control circuitry utilizes a MEMS comprising one or more members to change the first microdisplay from the second position to the third position.
- the 3D object may change locations (e.g., from the first location to the second location) during display of the piece of 3D content.
- the 3D object may be a first distance from a first viewpoint during the first segment of the piece of 3D content and may be a different distance from the first viewpoint during the second segment of the piece of 3D content.
- the control circuitry uses a first speed in which the 3D object changes locations within the 3D content to calculate a second speed at which one or more microdisplays are required to move from the second position to the third position.
- the control circuitry may use Equation 6 to determine the second speed at which to change the first microdisplay from the second position to the third position.
- control circuitry also changes the focal length of the lens to reduce the speed at which the control circuitry is required to move the first microdisplay from the second position to the third position to display the 3D object changing distances relative to the first viewpoint of the piece of 3D content.
- control circuitry manipulates the display of the piece of 3D content at step 1318 , so that the distance between the viewpoint and the 3D object in the piece of 3D content is larger. The larger distance between the viewpoint and the 3D object may reduce the speed at which the control circuitry is required to move the first microdisplay from the second position to the third position to display the 3D object changing distances relative to the first viewpoint of the piece of 3D content.
- control circuitry uses a combination of methodologies to reduce the speed at which the control circuitry is required to move the first microdisplay from the second position to the third position to display the 3D object changing distances relative to the first viewpoint of the piece of 3D content.
- control circuitry displays the second segment comprising the 3D object using the plurality of microdisplays.
- the third position of the first microdisplay is on a different plane relative to at least one microdisplay of the plurality of microdisplays.
- the first microdisplay may be a first vertical distance from a lens and a second microdisplay may be a second vertical distance from the lens.
- the control circuitry displaying the 3D object on the first microdisplay at the third position provides depth and reduces or eliminates the vergence-accommodation conflict.
- FIGS. 12 - 13 may be used with other suitable embodiments of this disclosure.
- some suitable steps and descriptions described in relation to FIGS. 12 - 13 may be implemented in alternative orders or in parallel to further the purposes of this disclosure.
- some suitable steps may be performed in any order or in parallel or substantially simultaneously to reduce lag or increase the speed of the system or method.
- Some suitable steps may also be skipped or omitted from the process.
- some suitable devices or equipment discussed in relation to FIGS. 1 A- 11 could be used to perform one or more of the steps in FIGS. 12 - 13 .
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
Where,
f: focal length of the lens 402.
d: distance between lens 402 and a microdisplay.
z: distance between the viewpoint and an object in the piece of 3D content.
Where,
f: focal length of the lens 402.
d1: distance 414 a between lens 402 and the first microdisplay 406 a.
z1: distance 412 a between the viewpoint and the first object 410 a in the piece of 3D content.
Where,
f: focal length of the lens 402.
d2: distance 414 b between lens 402 and the second microdisplay 406 b.
z2: distance 412 b between the viewpoint and the second object 410 b in the piece of 3D content.
Where,
f: focal length of the lens 402.
d3: distance 414 c between lens 402 and the third microdisplay 406 c.
z3: distance 412 c between the viewpoint and the third object 410 c in the piece of 3D content.
d1′: new distance between lens 402 and the first microdisplay 406 a.
z1: distance 412 a between the viewpoint and the first object 410 a in the piece of 3D content.
S: distance per second in which the first object 410 a changes from the first distance 412 a from the first viewpoint to the updated distance from the first viewpoint.
Equation (5) and Equation (2) may be used to find a second speed corresponding to the change of the first microdisplay 406 a from the distance 414 a between the first microdisplay 406 a and the lens 402 to the new distance between the first microdisplay 406 a and the lens 402. For example:
Sd: distance per second in which the first microdisplay 406 a changes from the distance 414 a between the first microdisplay 406 a and the lens 402 to the new distance (d1′) between the first microdisplay 406 a and the lens 402.
In some embodiments, the device 400 changes the focal length of the lens 402 to reduce the speed at which the device 400 would have to move the first microdisplay 406 a to display the first object 410 a changing distances relative to the first viewpoint of the piece of 3D content. In some embodiments, the device 400 manipulates the display of the piece of 3D content so that the distance 412 a between the viewpoint and the first object 410 a in the piece of 3D content is larger. The larger distance between the viewpoint and the first object 410 a reduces the speed at which the first microdisplay 406 a would have to move the first microdisplay 406 a to display the first object 410 a changing distances relative to the first viewpoint of the piece of 3D content. In some embodiments, the device uses a combination of methodologies to reduce the speed at which the first microdisplay 406 a would have to move to display the first object 410 a changing distances relative to the first viewpoint of the piece of 3D content.
Claims (20)
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US18/204,599 US12445588B2 (en) | 2023-06-01 | 2023-06-01 | Adaptive displays to address the vergence-accommodation conflict |
| US19/331,404 US20260019551A1 (en) | 2023-06-01 | 2025-09-17 | Adaptive displays to address the vergence-accommodation conflict |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US18/204,599 US12445588B2 (en) | 2023-06-01 | 2023-06-01 | Adaptive displays to address the vergence-accommodation conflict |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US19/331,404 Continuation US20260019551A1 (en) | 2023-06-01 | 2025-09-17 | Adaptive displays to address the vergence-accommodation conflict |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20240406367A1 US20240406367A1 (en) | 2024-12-05 |
| US12445588B2 true US12445588B2 (en) | 2025-10-14 |
Family
ID=93651841
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US18/204,599 Active 2043-10-22 US12445588B2 (en) | 2023-06-01 | 2023-06-01 | Adaptive displays to address the vergence-accommodation conflict |
| US19/331,404 Pending US20260019551A1 (en) | 2023-06-01 | 2025-09-17 | Adaptive displays to address the vergence-accommodation conflict |
Family Applications After (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US19/331,404 Pending US20260019551A1 (en) | 2023-06-01 | 2025-09-17 | Adaptive displays to address the vergence-accommodation conflict |
Country Status (1)
| Country | Link |
|---|---|
| US (2) | US12445588B2 (en) |
Citations (31)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20010005285A1 (en) * | 1999-12-28 | 2001-06-28 | Rohm Co., Ltd. | Head mounted display |
| US20070242237A1 (en) * | 2006-04-17 | 2007-10-18 | Thomas Clarence E | System and Methods for Angular Slice True 3-D Display |
| US20080024391A1 (en) * | 2006-07-31 | 2008-01-31 | Manuel Oliver | Image alignment method for binocular eyewear displays |
| US20080204548A1 (en) * | 2006-10-27 | 2008-08-28 | Emine Goulanian | Switchable optical imaging system and related 3d/2d image switchable apparatus |
| US20090005961A1 (en) * | 2004-06-03 | 2009-01-01 | Making Virtual Solid, L.L.C. | En-Route Navigation Display Method and Apparatus Using Head-Up Display |
| US20090156128A1 (en) * | 2007-12-12 | 2009-06-18 | Motorola, Inc. | Eyewear communications system |
| US20100060551A1 (en) * | 2007-09-26 | 2010-03-11 | Keiji Sugiyama | Beam scanning-type display device, method, program and integrated circuit |
| US20100214537A1 (en) * | 2009-02-23 | 2010-08-26 | Thomas Clarence E | System and Methods for Angular Slice True 3-D Display |
| US20120113092A1 (en) * | 2010-11-08 | 2012-05-10 | Avi Bar-Zeev | Automatic variable virtual focus for augmented reality displays |
| US20130002813A1 (en) * | 2011-06-29 | 2013-01-03 | Vaught Benjamin I | Viewing windows for video streams |
| US20140340389A1 (en) * | 2013-05-17 | 2014-11-20 | Nvidia Corporation | System, method, and computer program product to produce images for a near-eye light field display |
| US20150277129A1 (en) * | 2012-10-18 | 2015-10-01 | The Arizona Board Of Regents On Behalf Of The University Of Arizona | Stereoscopic displays with addressable focus cues |
| US20150301360A1 (en) * | 2012-12-13 | 2015-10-22 | Kopin Corporation | Spectacle with invisible optics |
| US20160003636A1 (en) * | 2013-03-15 | 2016-01-07 | Honda Motor Co., Ltd. | Multi-level navigation monitoring and control |
| US9268139B2 (en) * | 2013-03-15 | 2016-02-23 | Immy Inc. | Head mounted display with micro-display alignment mechanism |
| US20170078652A1 (en) * | 2014-03-05 | 2017-03-16 | The Arizona Board Of Regents On Behalf Of The University Of Arizona | A wearable 3d augmented reality display |
| US20180211448A1 (en) * | 2010-09-21 | 2018-07-26 | Telefonaktiebolaget Lm Ericsson (Publ) | Opacity filter for display device |
| US20180249086A1 (en) * | 2017-02-27 | 2018-08-30 | Seiko Epson Corporation | Display device and control method for display device |
| US20180321496A1 (en) * | 2017-05-03 | 2018-11-08 | Microsoft Technology Licensing, Llc | Near-to-eye display with steerable phased arrays |
| US20190179409A1 (en) * | 2017-12-03 | 2019-06-13 | Frank Jones | Enhancing the performance of near-to-eye vision systems |
| US20200278544A1 (en) * | 2019-02-28 | 2020-09-03 | Microsoft Technology Licensing, Llc | Active display alignment for multi-display device |
| US20210014473A1 (en) * | 2018-03-22 | 2021-01-14 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Methods of rendering light field images for integral-imaging-based light field display |
| US20210127046A1 (en) * | 2019-10-24 | 2021-04-29 | Sony Corporation | Device, method and computer program |
| US20210133994A1 (en) * | 2018-01-19 | 2021-05-06 | Pcms Holdings, Inc. | Multi-focal planes with varying positions |
| US20210165222A1 (en) * | 2017-12-21 | 2021-06-03 | Nokia Technologies Oy | Display Apparatus and Method |
| US20210181533A1 (en) * | 2019-12-11 | 2021-06-17 | Nvidia Corporation | Modular prescription augmented reality display |
| US20210255462A1 (en) * | 2019-12-20 | 2021-08-19 | Tobii Ab | Method, Computer Program Product and Binocular Headset Controller |
| US20220003990A1 (en) * | 2019-03-20 | 2022-01-06 | Sony Group Corporation | Image display apparatus |
| US20220075199A1 (en) * | 2018-12-28 | 2022-03-10 | Magic Leap, Inc. | Variable pixel density display system with mechanically-actuated image projector |
| US20240013752A1 (en) * | 2020-08-03 | 2024-01-11 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Perceptual-driven foveated displays |
| US11961431B2 (en) * | 2018-07-03 | 2024-04-16 | Google Llc | Display processing circuitry |
-
2023
- 2023-06-01 US US18/204,599 patent/US12445588B2/en active Active
-
2025
- 2025-09-17 US US19/331,404 patent/US20260019551A1/en active Pending
Patent Citations (31)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20010005285A1 (en) * | 1999-12-28 | 2001-06-28 | Rohm Co., Ltd. | Head mounted display |
| US20090005961A1 (en) * | 2004-06-03 | 2009-01-01 | Making Virtual Solid, L.L.C. | En-Route Navigation Display Method and Apparatus Using Head-Up Display |
| US20070242237A1 (en) * | 2006-04-17 | 2007-10-18 | Thomas Clarence E | System and Methods for Angular Slice True 3-D Display |
| US20080024391A1 (en) * | 2006-07-31 | 2008-01-31 | Manuel Oliver | Image alignment method for binocular eyewear displays |
| US20080204548A1 (en) * | 2006-10-27 | 2008-08-28 | Emine Goulanian | Switchable optical imaging system and related 3d/2d image switchable apparatus |
| US20100060551A1 (en) * | 2007-09-26 | 2010-03-11 | Keiji Sugiyama | Beam scanning-type display device, method, program and integrated circuit |
| US20090156128A1 (en) * | 2007-12-12 | 2009-06-18 | Motorola, Inc. | Eyewear communications system |
| US20100214537A1 (en) * | 2009-02-23 | 2010-08-26 | Thomas Clarence E | System and Methods for Angular Slice True 3-D Display |
| US20180211448A1 (en) * | 2010-09-21 | 2018-07-26 | Telefonaktiebolaget Lm Ericsson (Publ) | Opacity filter for display device |
| US20120113092A1 (en) * | 2010-11-08 | 2012-05-10 | Avi Bar-Zeev | Automatic variable virtual focus for augmented reality displays |
| US20130002813A1 (en) * | 2011-06-29 | 2013-01-03 | Vaught Benjamin I | Viewing windows for video streams |
| US20150277129A1 (en) * | 2012-10-18 | 2015-10-01 | The Arizona Board Of Regents On Behalf Of The University Of Arizona | Stereoscopic displays with addressable focus cues |
| US20150301360A1 (en) * | 2012-12-13 | 2015-10-22 | Kopin Corporation | Spectacle with invisible optics |
| US9268139B2 (en) * | 2013-03-15 | 2016-02-23 | Immy Inc. | Head mounted display with micro-display alignment mechanism |
| US20160003636A1 (en) * | 2013-03-15 | 2016-01-07 | Honda Motor Co., Ltd. | Multi-level navigation monitoring and control |
| US20140340389A1 (en) * | 2013-05-17 | 2014-11-20 | Nvidia Corporation | System, method, and computer program product to produce images for a near-eye light field display |
| US20170078652A1 (en) * | 2014-03-05 | 2017-03-16 | The Arizona Board Of Regents On Behalf Of The University Of Arizona | A wearable 3d augmented reality display |
| US20180249086A1 (en) * | 2017-02-27 | 2018-08-30 | Seiko Epson Corporation | Display device and control method for display device |
| US20180321496A1 (en) * | 2017-05-03 | 2018-11-08 | Microsoft Technology Licensing, Llc | Near-to-eye display with steerable phased arrays |
| US20190179409A1 (en) * | 2017-12-03 | 2019-06-13 | Frank Jones | Enhancing the performance of near-to-eye vision systems |
| US20210165222A1 (en) * | 2017-12-21 | 2021-06-03 | Nokia Technologies Oy | Display Apparatus and Method |
| US20210133994A1 (en) * | 2018-01-19 | 2021-05-06 | Pcms Holdings, Inc. | Multi-focal planes with varying positions |
| US20210014473A1 (en) * | 2018-03-22 | 2021-01-14 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Methods of rendering light field images for integral-imaging-based light field display |
| US11961431B2 (en) * | 2018-07-03 | 2024-04-16 | Google Llc | Display processing circuitry |
| US20220075199A1 (en) * | 2018-12-28 | 2022-03-10 | Magic Leap, Inc. | Variable pixel density display system with mechanically-actuated image projector |
| US20200278544A1 (en) * | 2019-02-28 | 2020-09-03 | Microsoft Technology Licensing, Llc | Active display alignment for multi-display device |
| US20220003990A1 (en) * | 2019-03-20 | 2022-01-06 | Sony Group Corporation | Image display apparatus |
| US20210127046A1 (en) * | 2019-10-24 | 2021-04-29 | Sony Corporation | Device, method and computer program |
| US20210181533A1 (en) * | 2019-12-11 | 2021-06-17 | Nvidia Corporation | Modular prescription augmented reality display |
| US20210255462A1 (en) * | 2019-12-20 | 2021-08-19 | Tobii Ab | Method, Computer Program Product and Binocular Headset Controller |
| US20240013752A1 (en) * | 2020-08-03 | 2024-01-11 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Perceptual-driven foveated displays |
Non-Patent Citations (27)
| Title |
|---|
| "Flexible display," Wikipedia, Retrieved from https://en.wikipedia.org/wiki/Flexible_display., Retrieved on Mar. 21, 2025, 1 page. |
| "Luminous Fabric, Glowing Fiber Optic Fabric, LED light up Cloth, Color Changeable Material," Retrieved from https://www.etsy.com/ca/listing/1283669248/luminous-fabric-glowing-fiber-optic, Dec. 2022. |
| Akeley K. et al., "A stereo display prototype with multiple focal distances," ACM transactions on graphics (TOG), vol. 23, Issue 3, 2004, pp. 804-813. |
| Antipa. N. et al., "Single-shot diffuser-encoded light field imaging," 2016 IEEE International Conference on Computational Photography (ICCP). IEEE, 2016, 11 pages. |
| Banks. MS. et al., "Consequences of incorrect focus cues in stereo displays," Journal of the Society for Information Display, vol. 24, Issue 7, 2008, 11 pages. |
| Gosline. AH. et al., "Percutaneous intracardiac beating-heart surgery using metal MEMS tissue approximation tools," The International journal of robotics research, vol. 31, Issue 9, 2012, pp. 1081-1093. |
| Huang. FC. et al., "The light field stereoscope," SIGGRAPH emerging technologies, 2015, 12 pages. |
| Jiang. F. et al., "A flexible MEMS technology and its first application to shear stress sensor skin," Proceedings IEEE The Tenth Annual International Workshop on Micro Electro Mechanical Systems. An Investigation of Micro Structures, Sensors, Actuators, Machines and Robots. IEEE, 1997, 6 pages. |
| Kamal. MH. et al., "Tensor low-rank and sparse light field photography," Computer Vision and Image Understanding, vol. 145, 2016, pp. 172-181. |
| Kraminda. G., "Resolving the vergence-accommodation conflict in head-mounted displays," IEEE transactions on visualization and computer graphics, vol. 22, Issue 7, 2015, pp. 1912-1931. |
| Lanman. D. et al., "Near-eye light field displays," ACM transactions on graphics (TOG), vol. 32, Issue 6, 2013, pp. 1-10. |
| Levoy, M., et al., "Light field rendering," SIGGRAPH96, Proceedings of the 23rd annual conference on Conputer Graphics Proceeding, 1996, pp. 31-42. |
| Li. B. et al., "A New MEMS Engine for Large-volume Subcutaneous Injectors," Stroke, 2022, 6 pages. |
| Love. GD. et al., "High-speed switchable lens enables the development of a volumetric stereoscopic display," Optics express, vol. 17, Issue 18, 2009, pp. 15716-15725. |
| Marinan. A. et al., "Payload characterization for CubeSat demonstration of MEMS deformable mirrors," Adaptive Optics Systems IV., vol. 9148, 2014, 17 pages. |
| Marwah. K. et al., "Compressive light field photography using overcomplete dictionaries and optimized projections," ACM Transactions on Graphics (TOG), vol. 32, Issue 4, 2013, pp. 1-12. |
| Padmanaban. N. et al., "Optimizing virtual reality for all users through gaze-contingent and adaptive focus displays," Proceedings of the National Academy of Sciences, vol. 114, Issue 9, 2017, pp. 2183-2188. |
| Rolland. JP. et al., "Multifocal planes head-mounted displays," Applied optics, vol. 39, Issue 19, 2000, pp. 3209-3215. |
| Severson. SA. et al., "KAPAO: a MEMS-based natural guide star adaptive optics system," MEMS Adaptive Optics VII., vol. 8617, 2013, 10 pages. |
| Shibata. T. et al., "The zone of comfort: Predicting visual discomfort with stereo displays," Journal of vision, vol. 11, Issue 8, 2011, 29 pages. |
| Teague. MR., "Deterministic phase retrieval: a Green's function solution," Journal of the Optical Society of America, vol. 73, Issue 11, 1983, pp. 1434-1441. |
| Veeraraghavan, A. et al., "Dappled photography: Mask enhanced cameras for heterodyned light fields and coded aperture refocusing," ACM Trans. Graph., vol. 26, Issue 3, 2007, 14 pages. |
| Wei. LY. et al., "Improving light field camera sample design with irregularity and aberration," ACM Transactions on Graphics (TOG), vol. 34, Issue 4, 2015, pp. 1-11. |
| Xianfeng. X. et al., "Design and realization of control system of a novel digital hydraulic 2-DoF motion platform," 1st International Conference on Mechanical Engineering and Material Science (MEMS 2012). Atlantis Press, 2012, 4 pages. |
| Yamamoto. T. et al., "LIFLET: Light field live with thousands of lenslets," ACM SIGGRAPH 2004 Emerging technologies. 2004. |
| Yunas. J. et al., "Polymer-based MEMS electromagnetic actuator for biomedical application: A review," Polymers, vol. 2, Issue 5, 2020, 21 pages. |
| Zhang. W. et al., "Development of miniature camera module integrated with solid tunable lens driven by MEMS-thermal actuator," Journal of Microelectromechanical Systems, vol. 26, Issue 1, 2016, pp. 84-94. |
Also Published As
| Publication number | Publication date |
|---|---|
| US20260019551A1 (en) | 2026-01-15 |
| US20240406367A1 (en) | 2024-12-05 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| KR101789357B1 (en) | Automatic focus improvement for augmented reality displays | |
| EP3000020B1 (en) | Hologram anchoring and dynamic positioning | |
| EP3855290B1 (en) | Remote rendering for virtual images | |
| EP3528097B1 (en) | Hybrid world/body locked hud on an hmd | |
| KR101912958B1 (en) | Automatic variable virtual focus for augmented reality displays | |
| US9934614B2 (en) | Fixed size augmented reality objects | |
| US9329682B2 (en) | Multi-step virtual object selection | |
| KR101960980B1 (en) | Optimized focal area for augmented reality displays | |
| US20140368537A1 (en) | Shared and private holographic objects | |
| US10134174B2 (en) | Texture mapping with render-baked animation | |
| WO2016118344A1 (en) | Fixed size augmented reality objects | |
| US20210373654A1 (en) | Determining angular acceleration | |
| US12445588B2 (en) | Adaptive displays to address the vergence-accommodation conflict | |
| US12008720B1 (en) | Scene graph assisted navigation | |
| US20250182310A1 (en) | Foveated imaging based on machine learning modeling of depth mapping | |
| WO2025122671A1 (en) | Foveated imaging based on machine learning modeling of depth mapping | |
| CN121127815A (en) | Adjustable optical field for augmented reality |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: ADEIA GUIDES INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BASU, ANUP;LAL, DHANANJAY;XU, NING;SIGNING DATES FROM 20230602 TO 20230605;REEL/FRAME:064625/0063 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |