US20140292812A1 - Visual training devices, systems, and methods - Google Patents
Visual training devices, systems, and methods Download PDFInfo
- Publication number
- US20140292812A1 US20140292812A1 US14/300,017 US201414300017A US2014292812A1 US 20140292812 A1 US20140292812 A1 US 20140292812A1 US 201414300017 A US201414300017 A US 201414300017A US 2014292812 A1 US2014292812 A1 US 2014292812A1
- Authority
- US
- United States
- Prior art keywords
- course
- data
- image
- user
- eyewear article
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
- A63B71/0619—Displays, user interfaces and indicating devices, specially adapted for sport equipment, e.g. display mounted on treadmills
- A63B71/0622—Visual, audio or audio-visual systems for entertaining, instructing or motivating the user
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
- A63B71/0686—Timers, rhythm indicators or pacing apparatus using electric or electronic means
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J1/00—Photometry, e.g. photographic exposure meter
- G01J1/02—Details
- G01J1/04—Optical or mechanical part supplementary adjustable parts
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J1/00—Photometry, e.g. photographic exposure meter
- G01J1/02—Details
- G01J1/04—Optical or mechanical part supplementary adjustable parts
- G01J1/0407—Optical elements not provided otherwise, e.g. manifolds, windows, holograms, gratings
- G01J1/0414—Optical elements not provided otherwise, e.g. manifolds, windows, holograms, gratings using plane or convex mirrors, parallel phase plates, or plane beam-splitters
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J1/00—Photometry, e.g. photographic exposure meter
- G01J1/42—Photometry, e.g. photographic exposure meter using electric radiation detectors
- G01J1/4228—Photometry, e.g. photographic exposure meter using electric radiation detectors arrangements with two or more detectors, e.g. for sensitivity compensation
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C11/00—Non-optical adjuncts; Attachment thereof
- G02C11/10—Electronic devices other than hearing aids
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03F—PHOTOMECHANICAL PRODUCTION OF TEXTURED OR PATTERNED SURFACES, e.g. FOR PRINTING, FOR PROCESSING OF SEMICONDUCTOR DEVICES; MATERIALS THEREFOR; ORIGINALS THEREFOR; APPARATUS SPECIALLY ADAPTED THEREFOR
- G03F7/00—Photomechanical, e.g. photolithographic, production of textured or patterned surfaces, e.g. printing surfaces; Materials therefor, e.g. comprising photoresists; Apparatus specially adapted therefor
- G03F7/70—Microphotolithographic exposure; Apparatus therefor
- G03F7/70058—Mask illumination systems
- G03F7/70091—Illumination settings, i.e. intensity distribution in the pupil plane or angular distribution in the field plane; On-axis or off-axis settings, e.g. annular, dipole or quadrupole settings; Partial coherence control, i.e. sigma or numerical aperture [NA]
- G03F7/70116—Off-axis setting using a programmable means, e.g. liquid crystal display [LCD], digital micromirror device [DMD] or pupil facets
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03F—PHOTOMECHANICAL PRODUCTION OF TEXTURED OR PATTERNED SURFACES, e.g. FOR PRINTING, FOR PROCESSING OF SEMICONDUCTOR DEVICES; MATERIALS THEREFOR; ORIGINALS THEREFOR; APPARATUS SPECIALLY ADAPTED THEREFOR
- G03F7/00—Photomechanical, e.g. photolithographic, production of textured or patterned surfaces, e.g. printing surfaces; Materials therefor, e.g. comprising photoresists; Apparatus specially adapted therefor
- G03F7/70—Microphotolithographic exposure; Apparatus therefor
- G03F7/708—Construction of apparatus, e.g. environment aspects, hygiene aspects or materials
- G03F7/7085—Detection arrangement, e.g. detectors of apparatus alignment possibly mounted on wafers, exposure dose, photo-cleaning flux, stray light, thermal load
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03F—PHOTOMECHANICAL PRODUCTION OF TEXTURED OR PATTERNED SURFACES, e.g. FOR PRINTING, FOR PROCESSING OF SEMICONDUCTOR DEVICES; MATERIALS THEREFOR; ORIGINALS THEREFOR; APPARATUS SPECIALLY ADAPTED THEREFOR
- G03F7/00—Photomechanical, e.g. photolithographic, production of textured or patterned surfaces, e.g. printing surfaces; Materials therefor, e.g. comprising photoresists; Apparatus specially adapted therefor
- G03F7/70—Microphotolithographic exposure; Apparatus therefor
- G03F7/708—Construction of apparatus, e.g. environment aspects, hygiene aspects or materials
- G03F7/70858—Environment aspects, e.g. pressure of beam-path gas, temperature
- G03F7/70883—Environment aspects, e.g. pressure of beam-path gas, temperature of optical system
- G03F7/70891—Temperature
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/163—Wearable computers, e.g. on a belt
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/003—Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
- G09G5/005—Adapting incoming signals to the display format of the display terminal
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/36—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
- G09G5/37—Details of the operation on graphic patterns
- G09G5/377—Details of the operation on graphic patterns for mixing or overlaying two or more graphic patterns
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B24/00—Electric or electronic controls for exercising apparatus of preceding groups; Controlling or monitoring of exercises, sportive games, training or athletic performances
- A63B24/0003—Analysing the course of a movement or motion sequences during an exercise or trainings sequence, e.g. swing for golf or tennis
- A63B24/0006—Computerised comparison for qualitative assessment of motion sequences or the course of a movement
- A63B2024/0009—Computerised real time comparison with previous movements or motion sequences of the user
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B24/00—Electric or electronic controls for exercising apparatus of preceding groups; Controlling or monitoring of exercises, sportive games, training or athletic performances
- A63B24/0003—Analysing the course of a movement or motion sequences during an exercise or trainings sequence, e.g. swing for golf or tennis
- A63B24/0006—Computerised comparison for qualitative assessment of motion sequences or the course of a movement
- A63B2024/0012—Comparing movements or motion sequences with a registered reference
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
- A63B71/0619—Displays, user interfaces and indicating devices, specially adapted for sport equipment, e.g. display mounted on treadmills
- A63B2071/0658—Position or arrangement of display
- A63B2071/0661—Position or arrangement of display arranged on the user
- A63B2071/0666—Position or arrangement of display arranged on the user worn on the head or face, e.g. combined with goggles or glasses
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
- A63B2071/0691—Maps, e.g. yardage maps or electronic maps
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2220/00—Measuring of physical parameters relating to sporting activity
- A63B2220/10—Positions
- A63B2220/12—Absolute positions, e.g. by using GPS
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2220/00—Measuring of physical parameters relating to sporting activity
- A63B2220/10—Positions
- A63B2220/16—Angular positions
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2220/00—Measuring of physical parameters relating to sporting activity
- A63B2220/17—Counting, e.g. counting periodical movements, revolutions or cycles, or including further data processing to determine distances or speed
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2220/00—Measuring of physical parameters relating to sporting activity
- A63B2220/30—Speed
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2220/00—Measuring of physical parameters relating to sporting activity
- A63B2220/40—Acceleration
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2220/00—Measuring of physical parameters relating to sporting activity
- A63B2220/62—Time or time measurement used for time reference, time stamp, master time or clock signal
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2220/00—Measuring of physical parameters relating to sporting activity
- A63B2220/80—Special sensors, transducers or devices therefor
- A63B2220/806—Video cameras
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2225/00—Miscellaneous features of sport apparatus, devices or equipment
- A63B2225/20—Miscellaneous features of sport apparatus, devices or equipment with means for remote communication, e.g. internet or the like
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2225/00—Miscellaneous features of sport apparatus, devices or equipment
- A63B2225/50—Wireless data transmission, e.g. by radio transmitters or telemetry
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B69/00—Training appliances or apparatus for special sports
- A63B69/0028—Training appliances or apparatus for special sports for running, jogging or speed-walking
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B69/00—Training appliances or apparatus for special sports
- A63B69/16—Training appliances or apparatus for special sports for cycling, i.e. arrangements on or for real bicycles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S19/00—Satellite radio beacon positioning systems; Determining position, velocity or attitude using signals transmitted by such systems
- G01S19/01—Satellite radio beacon positioning systems transmitting time-stamped messages, e.g. GPS [Global Positioning System], GLONASS [Global Orbiting Navigation Satellite System] or GALILEO
- G01S19/13—Receivers
- G01S19/14—Receivers specially adapted for specific applications
- G01S19/19—Sporting applications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2370/00—Aspects of data communication
- G09G2370/12—Use of DVI or HDMI protocol in interfaces along the display data pipeline
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2380/00—Specific applications
Definitions
- the present disclosure relates generally to visual training devices, systems, and methods, collectively visual training aids.
- location aware visual training aids are described herein.
- Training assistance devices in the art are not entirely satisfactory. Specifically, training aids in the art provide users with limited training data during a workout. Training data typically provided by conventional training devices includes current and average speed information, pace information, route information, the time in which the user completes a particular route, current and average heart rate information, and current or average cadence information.
- Training assistance devices in the art that provide visual data often require users to look away from their path to a separate device, such as a watch or mobile device. Both looking away from the path and handling other devices can have detrimental effect on an athlete's pace and route. Additionally, these distractions may cause safety concerns due to the lack of focus on the route.
- video eyewear technologies now allows users to view computer generated graphics on a display proximate their eyes and substantially cover their entire field of vision.
- Such technologies generally achieve this functionality by attaching projectors to a set of eyewear that are capable of projecting graphical images on the lenses.
- These displays may receive a computer generated image from an external source and use a series of mirrors to project this image on the lens, thereby placing computer generated graphical images in the user's natural line of vision.
- technologies by which images are displayed on the lenses of such video eyewear including more traditional display technologies such as LCD.
- Augmented reality displays comprise a computer generated graphical display laid over a portion of a user's or mobile device's natural field of vision. These “augmented reality” displays allow a user to view computer generated images including data related to objects in their natural field of vision. Augmented reality displays may include any display including both natural and computer generated elements.
- Examples of mobile computing devices include Apple iPhone mobile devices, mobile devices designed with the Android operating system, other mobile phone devices, portable media players, personal computers, watches, and other mobile computing devices that are capable of executing third party developed software.
- the present disclosure is directed to visual training aids including an eyewear article including a lens, an image generator mounted to the eyewear article in a position to display an image on the lens, a processor operatively connected to and in data communication with the image generator, a global positioning system operatively connected to and in data communication with the processor, a computer readable medium operatively connected to and in data communication with the processor.
- the visual training aid includes a pair of eyeglasses.
- the visual training aid includes a display monitor mounted to the eyewear.
- the visual training aid includes a camera mounted to the eyewear.
- FIG. 1 is a schematic view of a computing device that may be used to implement various aspects of visual training aids.
- FIG. 2 is a schematic view of a first example of a visual training aid.
- FIG. 3 is a rear elevation view of a visual training aid including a pair of eyeglasses.
- FIG. 4 is a side elevation view of the visual training aid shown in FIG. 3 .
- FIG. 5 illustrates an example of a visual training aid with a forward facing camera.
- FIG. 6 illustrates an example of a visual training aid with a forward facing camera and a rearward facing camera and where components of the visual training aid are connected by a wireless data communications link.
- visual training aids may be implemented using electronic circuitry configured to perform one or more functions.
- the visual training aid may be implemented using one or more application-specific integrated circuits (ASICs). More typically, however, components of various examples of the invention will be implemented using a programmable computing device executing firmware or software instructions, or by some combination of purpose-specific electronic circuitry and firmware or software instructions executing on a programmable computing device.
- ASICs application-specific integrated circuits
- FIG. 1 shows one illustrative example of a computer 101 that can be used to implement various embodiments of the invention.
- Computer 101 may be incorporated within a variety of consumer electronic devices, such as personal media players, cellular phones, smart phones, personal data assistants, global positioning system devices, and the like.
- Computer 101 has a computing unit 103 .
- Computing unit 103 typically includes a processing unit 105 and a system memory 107 .
- Processing unit 105 may be any type of processing device for executing software instructions, but will conventionally be a microprocessor device.
- System memory 107 may include both a read-only memory (ROM) 109 and a random access memory (RAM) 111 .
- ROM read-only memory
- RAM random access memory
- both read-only memory (ROM) 109 and random access memory (RAM) 111 may store software instructions to be executed by processing unit 105 .
- Processing unit 105 and system memory 107 are connected, either directly or indirectly, through a bus 113 or alternate communication structure to one or more peripheral devices.
- processing unit 105 or system memory 107 may be directly or indirectly connected to additional memory storage, such as a hard disk drive 115 , a removable magnetic disk drive 117 , an optical disk drive 119 , and a flash memory card 121 .
- Processing unit 105 and system memory 107 also may be directly or indirectly connected to one or more input devices 123 and one or more output devices 125 .
- Input devices 123 may include, for example, a keyboard, touch screen, a remote control pad, a pointing device (such as a mouse, touchpad, stylus, trackball, or joystick), a scanner, a camera or a microphone.
- Output devices 125 may include, for example, a monitor display, an integrated display, television, printer, stereo, or speakers.
- computing unit 103 will be directly or indirectly connected to one or more network interfaces 127 for communicating with a network.
- This type of network interface 127 also sometimes referred to as a network adapter or network interface card (NIC), translates data and control signals from computing unit 103 into network messages according to one or more communication protocols, such as the Transmission Control Protocol (TCP), the Internet Protocol (IP), and the User Datagram Protocol (UDP). These protocols are well known in the art, and thus will not be discussed here in more detail.
- An interface 127 may employ any suitable connection agent for connecting to a network, including, for example, a wireless transceiver, a power line adapter, a modem, or an Ethernet connection.
- the computing device may be connected to a variety of other peripheral devices, including some that may perform input, output and storage functions, or some combination thereof.
- the computer 101 may be connected to a digital music player, such as an IPOD® brand digital music player or IPHONE® brand smartphone available from Apple, Inc. of Cupertino, Calif.
- this type of digital music player can serve as both an output device for a computer (e.g., outputting music from a sound file or pictures from an image file) and a storage device.
- this type of digital music play also can serve as an output device for outputting visual training information, as will be discussed in more detail below.
- computer 101 may be connected to or otherwise include one or more other peripheral devices, such as a telephone.
- the telephone may be, for example, a wireless “smart phone.” As known in the art, this type of telephone communicates through a wireless network using radio frequency transmissions.
- a “smart phone” may also provide a user with one or more data management functions, such as sending, receiving and viewing electronic messages (e.g., electronic mail messages, SMS text messages, etc.), recording or playing back sound files, recording or playing back image files (e.g., still picture or moving video image files), viewing and editing files with text (e.g., Microsoft Word or Excel files, or Adobe Acrobat files), etc. Because of the data management capability of this type of telephone, a user may connect the telephone with computer 101 so that their data maintained may be synchronized.
- electronic messages e.g., electronic mail messages, SMS text messages, etc.
- recording or playing back sound files e.g., still picture or moving video image files
- viewing and editing files with text e.g
- peripheral devices may be included with our otherwise connected to a computer 101 of the type illustrated in FIG. 1 , as is well known in the art.
- a peripheral device may be permanently or semi-permanently connected to computing unit 103 .
- computing unit 103 hard disk drive 117 , removable optical disk drive 119 and a display are semi-permanently encased in a single housing.
- Computer 101 may include, for example, one or more communication ports through which a peripheral device can be connected to computing unit 103 (either directly or indirectly through bus 113 ). These communication ports may thus include a parallel bus port or a serial bus port, such as a serial bus port using the Universal Serial Bus (USB) standard or the IEEE 1394 High Speed Serial Bus standard (e.g., a Firewire port). Alternately or additionally, computer 101 may include a wireless data “port,” such as a Bluetooth interface, a Wi-Fi interface, an infrared data port, or the like.
- USB Universal Serial Bus
- IEEE 1394 High Speed Serial Bus standard e.g., a Firewire port
- computer 101 may include a wireless data “port,” such as a Bluetooth interface, a Wi-Fi interface, an infrared data port, or the like.
- a computing device employed according various examples of the invention may include more components than computer 101 illustrated in FIG. 1 , fewer components than computer 101 , or a different combination of components than computer 101 .
- Some implementations of the invention may employ one or more computing devices that are intended to have a very specific functionality, such as a digital music player or server computer. These computing devices may thus omit unnecessary peripherals, such as the network interface 115 , removable optical disk drive 119 , printers, scanners, external hard drives, etc.
- Some implementations of the invention may alternately or additionally employ computing devices that are intended to be capable of a wide variety of functions, such as a desktop or laptop personal computer. These computing devices may have any combination of peripheral devices or additional components as desired.
- a visual training aid 210 includes eyeglasses 220 , a processor 230 , memory 240 , a global positioning system 250 , and an image generator 260 .
- Visual training aid 210 further includes an optional image capturing device or camera or 280 to capture images of the terrain surrounding the user. Additional components are included in certain examples, such as laser range finders and radar systems. Appropriate clocks and other time keeping devices, whether physical or software based, are contemplated for measuring relevant time intervals. Additionally or alternatively, the visual training aid may include data cables, wireless data transmitters and receivers, and removable computer readable storage mediums.
- visual training aids such as visual training aid 210
- Visual training aids described herein may be used in a variety of activities, including running, skiing, snowboarding, bicycling, horseback riding, racecar driving, motorcycle driving, and other timed competitive endeavors.
- the visual data helps to provide the user with an awareness of his position relative to the position of a competitor or to his own prior performance on a course. This visual data may thus stimulate the user's competitive drive and enhance his workout performance and training satisfaction.
- the visual training aids enhance training effectiveness and motivation through simulating elements of competition during solo training endeavors.
- a visual training aid 310 shown in FIG. 3 for demonstrating an example of simulated competition provided by the visual training aids described herein, the reader can see that a user utilizing visual training aid 310 sees an augmented reality display 370 including a computer generated representation of a first person 372 and a computer generated image of a second person 374 .
- a person in this context may also be referred to as a “competitor,” a “virtual competitor,” or a “prior performer.”
- the computer generated images of the first and second persons may be either static or animated and may depict another person's performance on the course or the user's own past performance on the course.
- the computer generated images depicting the apparent position of the virtual competitors 372 and 374 relative to the wearer and the surroundings is continually updated based on the position of competitors 372 and 374 on the course during a prior training session relative the user's current position at the same given time interval.
- representations of the competitor are displayed such that they appear in front of the user from the user's current vantage point. If the user's performance is faster than the competitor's prior pace, representations of the competitor are displayed such that they appear behind the user from the user's vantage point. If a user's pace is substantially the same as the competitor's prior pace on the course, representations of the competitor are projected to appear to the side of the user.
- the relative position of the trainer and representation of the prior performer will be constantly changing as their respective velocities change throughout the course.
- Eyewear 220 serves to support selected components of visual training aid 210 and to position an augmented reality display in a position to be viewed by a user of the visual training aid.
- Eyewear may include monocles, goggles, head mounted displays, helmet mounted displays, or any other arrangement of lenses supported proximate the eye.
- the computer generated display of virtual competitors is displayed on a mobile computing device, such as the display or screen of a mobile phone, including an iPhone® brand mobile phone, a personal media player, including the iTouch® personal media player, a personal data assistant, and the like.
- the user carries the mobile device with him during a training session and looks at the screen, for example, by holding the mobile device in front of him, to see on the screen or display the position of virtual competitors relative to his own position.
- the augmented reality display on the mobile device includes a map depicting the course, similar to a global positioning system map depicting a route, with virtual competitors positioned relative to the user's current position on the map.
- the augmented reality display is a real-time image of the surroundings visible through a camera lens of the mobile device with computer generated virtual competitors represented on the display in a position reflecting the difference in position between the user's current position and the person's position on the course during a prior training session at the same given time interval.
- Processor 230 is programmed with instructions for receiving data from components of visual training aid 210 , including memory 240 , global positioning system 250 , camera 280 , and clock or timer 290 .
- Data from components of visual training aid 210 such as global positioning system 250 , camera 280 , and clock 290 may be in a raw format that is not directly usable by other components of visual training aid 210 , such as image generator 260 or memory 240 .
- processor is programmed with instructions for translating, computing, converting, correlating, and integrating (collectively “processing”) data from each component into usable formats.
- processor is programmed with instruction for processing prior position data and prior time interval data from memory 240 , current position data from global positioning system 250 , terrain image data from camera 280 , and current time interval data from clock 290 .
- the processed data may be stored into memory 240 , communicated to databases or devices over the internet, or sent to other components of visual training aid 210 , such as image generator 260 .
- processor 230 may process and store processed data, visual training aid 210 storing raw data into memory 240 or into other databases is equally within this disclosure.
- Data collected by components of visual training aid 210 may be displayed in a readable way, stored and retrieved to local or remote storage means, including mobile devices, cloud technologies, internet servers, and social networking applications and sites.
- Reading past position data from memory 240 may include reading multiple sets of past position data from memory 240 .
- each set of past position data may correspond to a person's position on the course during prior training sessions at a given time interval relative to the respective course start times for each prior training session.
- the past position data may correspond to different people's position or a single person's position during two or more prior training sessions.
- Processor 230 is programmed with instruction for determining a current position of the user or eyewear article on a course at a given time interval relative to a current course start time. To make these determinations, processor 230 dynamically current position data from global positioning system 260 and time interval data from clock 290 . Additionally or alternatively, the processor may receive one or more of these data from memory.
- Processor 230 is further programmed with instructions for generating competitor image data depicting a person's position on a course relative to the user's current position at a given time interval.
- the competitor image data may include data corresponding to images of multiple competitors.
- the competitor image data includes instructions for image generator 260 to display an image on the lens depicting a person's or competitor's position on the course during a prior training session relative to the current position of the user or of eyewear article 220 at a given time interval.
- the competitor image data may include various parameters determining how the image of the virtual competitor will be displayed.
- the competitor image data may include where in the user's field of view the computer generated image of the virtual competitor is to be displayed to account for where the user is currently looking.
- the competitor image data may include the size of the image of the virtual competitor to display to account for proper distance and perspective.
- the competitor image data may include the profile of the virtual competitor to display, e.g., a view of the virtual competitor's back, left side, right side, or front, to account for where the user is looking and the respective position of the user and the person's prior position on the course at the relevant time interval.
- processor 230 may be programmed with instructions to dynamically adjust the direction in which the graphical representation of a prior performer's position is displayed based on dynamically updated data representing the direction the user is currently looking. For example, if the user is looking straight ahead and the competitor's prior performance at that given time interval was ahead of the user's current position, the processor instructions specify that the graphical representation of the prior performer be displayed to appear in front of the user. If the user were to look to his right 60°, the processor instructions specify that the graphical representation of the prior performer be displayed to appear 600 to the left of the user.
- the visual training aids described herein may include any conventional means for detecting the direction a user is looking.
- the visual training aid may include eye tracking devices, accelerometers to detect head position and orientation, image capturing devices to capture images of the terrain around the wearer, laser range finders, radar, and computer vision technologies.
- Processor 230 is programmed with instructions for outputting or sending data to components of visual training aid 210 , including memory 240 and image generator 260 .
- processor 230 is programmed with instructions for sending current position data and current time interval data to memory 250 .
- processor 230 is programmed with instructions for sending competitor image data to image generator 260 , including competitor image data for multiple virtual competitors.
- Processor 230 may be any conventional processor suitable for inputting, processing, and outputting data consistent with the functions and programmed instructions described herein.
- the processor may include multiple processors programmed to carry out specific instructions in concert.
- the processor may be specially fabricated for use in the visual training aids described herein or may employ commercially available processors or processors used for other devices.
- the processor used by visual training aids may be a processor incorporated within an Apple iPhone brand mobile phone, which is designed for functions beyond those applicable to a visual training aid.
- User installed software may provide all necessary instructions and drivers to achieve the functions described herein of a visual training aid.
- Other suitable processors include those utilized by Apple iTouch® brand mobile devices, mobile devices designed with the Android® operating system, other mobile devices, including portable media players, personal computers, watches, and other mobile computing devices that are capable of executing third-party developed software.
- processor system may include operating system software.
- Memory 240 may comprise any conventional data storage medium known in the art.
- memory 240 may include magnetic media, optical media, flash media, and the like.
- Memory 240 may also include a plurality of memory units, each memory unit being the same media or one or more being different types of media.
- global positioning system 250 dynamically determines the current, real-time position of the user as he moves about a course. This data may be described as current position data. Additionally or alternatively, current position data may be obtained from external sensors that monitor the position of the user and/or a component of the visual training aid. Other technologies, including pedometers, accelerometers, and magnetometers, may work in conjunction with global positioning systems, and in some examples replace them, to generate position and other training data.
- Any conventional global positioning system may be used for generating current position data.
- commercially available, stand-alone global positioning systems as well as global positioning systems incorporated into other devices, such as cellular phones and portable media players, are suitable.
- Appropriate software, cables, or wireless data protocols may be employed to communicate the current position data from the global positioning system to the processor.
- the processor used by a global positioning system itself may be programmed to carry out the functions and instructions of the visual training aids described herein.
- the current position data is sent to processor 230 for generating instructions for use by image generator 260 .
- processor 230 uses the current position data and time stamped past position data of a prior person (or persons) stored in memory 240 to determine the position of the user relative to the person's position on the course during a prior training session at a given time interval.
- Time stamped position data refers to position data cross referenced with time interval data detailing how long into a training session person was at a given position.
- Processor 230 then sends image data to image generator 260 for displaying an image to the user depicting the person's position on the course during the prior training session relative to the current position of the user at the given time interval.
- Processor 230 may be configured to use the current position data in conjunction with timing data from clock 290 to generate a wide range of training data.
- the training data may include the user's current speed, average speed, cadence information, and workout duration among various other useful metrics relating to the current workout, including previous pace in which the same route was completed.
- the training data may also include cues for the user, including graphical cues assisting the user to follow a route, such as maps and arrows, and cues assisting the user in knowing when to begin a particular workout.
- the training data may be displayed to the user over the course of a workout and/or may be stored in memory 240 for later retrieval.
- processor 230 is configured to use the current position data in conjunction with timing data from clock 290 to generate and store a set of time-stamped position data for use in subsequent training sessions.
- the time stamped position data generated during a current training session can serve as past position data during subsequent training sessions, either by a 3 rd party trainee or by the user himself. That is, the user's time stamped position data processed by processor 230 and stored in memory 240 may be used by processor 230 during a subsequent training session as the basis for generating a virtual competitor for the user of visual training aid.
- Image generator 260 displays an image on the lens for viewing by the user of visual training aid 210 based on competitor image data supplied by processor 230 .
- the image generator is configured to display a series of images in succession to define a video or an animated depiction of the virtual competitor.
- Image generator 260 may include a wide variety of image display technologies.
- the image generator includes a projector configured to project an image or a series of images onto the lens.
- the image generator includes a hologram generating system for generating holograms viewable by the user.
- the image generator includes a heads-up-display projection system.
- the image generator is configured to display in image visible to the user on a screen or display that at least partially blocks a user's view through the lens.
- the visual training aid may include an opaque screen positioned across at least a portion of at least one of the lenses in a pair of eyeglasses and a projector configured to project an image onto the screen.
- the image generator includes a display monitor, such as a liquid crystal display, positioned across at least a portion of a lens for displaying an image or a series of images to a user. While the user is unable to see through the screen or the display monitor, the image generator may display images corresponding to the user's surroundings based on terrain image data supplied by 230 processor in conjunction with camera 280 .
- Camera 280 serves to capture images of a course, including the user's surroundings and the corresponding terrain.
- Data corresponding to images of the course and surroundings may be described as terrain image data or course image data.
- the terrain image data When displayed on the lens of eyewear 220 , the terrain image data enables the user to perceive his surroundings even if he does not have a direct line of sight to his surroundings.
- the terrain image data may enable the user to perceive his surroundings when an opaque screen or display monitor is positioned in front of one of more of his eyes.
- camera 280 is a video camera configured to capture a series of images in succession to define a video.
- a wide variety of acceptable cameras and/or video cameras are commercially available.
- the resolution of the cameras may be selected to optimize data throughput and image quality. Acceptable resolutions include interlaced and progressive image scans at 480 , 525 , 720 , 1080 vertical scanning lines, as well as vertical scanning lines in between, higher, and lower than these example values.
- the visual training aid includes two or more cameras.
- a first camera is positioned to capture forward course image data corresponding to images of the course forward of a user wearing the eyewear and a second camera is positioned to capture rearward course image data corresponding to images of the course behind the user.
- a sufficient number of cameras are used to capture images providing a substantially 360° view of the course.
- the camera is an optional feature of the visual training aids described herein.
- the visual training aid does not include a camera.
- the image generator is configured to display images of virtual competitors on a substantially transparent lens, the user can see the course and the terrain through the lens without the need for a camera.
- the depictions of the virtual competitors are integrated with the actual view of the course and the terrain through the lens by where and how they are displayed on the lens.
- Clock 290 may be any conventional mechanical or software based time keeping device.
- a clock may serve in conjunction with a processor in a computing device, clock 290 serves as a timer for generating time interval data.
- the time interval data corresponds to the time difference between a given time on a course after starting a training session on the course.
- a current time interval is applicable to a user's current training session and is defined as the user's current time on a course minus the user's current course start time.
- a past time interval is applicable to a person's prior training session on a course and is defined as the time difference between a given time on the course and the person's past course start time.
- Clock 290 is configured to dynamically output time interval data to processor 230 .
- Processor 230 uses the current time interval data to determine which past position data is applicable for generating image data to image generator 260 .
- processor 230 will select past position data having a past time interval equal to the current time interval to correlate a person's position on the course during a prior training session with a user's current position on a course at a same given time interval. In this manner, the relative position of the virtual competitor at a given time interval corresponding to how long the user has currently been on the course can be displayed to the user.
- Visual training aid 310 includes many similar or identical features to visual training aid 210 combined in unique and distinct ways. Thus, for the sake of brevity, each feature of visual training aid 310 will not be redundantly explained. Rather, key distinctions between visual training aid 210 and visual training aid 310 will be described in detail and the reader should reference the discussion above for features substantially similar between the two visual training aids.
- visual training aid 310 includes an eyewear article 320 in the form of eyeglasses 321 , a processor 330 , memory 340 , a global positioning system 350 , a first image generator 360 , a second image generator 362 , and a clock 390 .
- visual training aid 310 generates an augmented reality display 370 to serve as a training aid for a user.
- all components of visual training aid 310 is supported by eyewear 320 .
- Eyeglasses 321 supports components of visual training aid 310 and positions augmented reality display 370 in a position to be viewed by a user.
- eyeglasses 321 include a first lens 322 , a second lens 324 , and a frame 338 .
- the eyewear article includes a single lens.
- First and second lenses 322 and 324 are made of a substrate that is generally suitable for the display of an image on their surface for transmission of light through the lens.
- one or both of the lenses are configured to transmit light primarily in a first direction towards the user's eye, with reduced or no transmission of light in the opposite direction.
- a wide range of conventional lenses may be used.
- Frame 327 supports a first lens support 328 and a second lens support 329 .
- Frame 327 may be made from any material conventionally used to construct eyeglass frames, including plastics and metals.
- the frame may adopt a wide variety of conventional eyewear frame designs.
- frame 327 is specially designed with a compartment 325 for housing components of visual training aid 310 , including processor 330 , memory 340 , global positioning system 350 , and clock 390 . Additional components, such as batteries, may be housed in compartment 325 as well.
- Visual training aid 310 can be operated in multiple modes.
- modes of operation include a mode where both first image generator 360 and second image generator 362 are displaying augmented reality displays, a mode where just one of the first and second image generators are displaying an augmented reality display, and a mode where neither the first and second image generators are displaying an augmented reality display.
- the mode of operation depicted in FIG. 3 is a mode where only first image generator 360 is displaying an augmented reality display 370 on only first lens 322 .
- visual training aid 310 may be operated in a mode where second image generator 362 is displaying an augmented reality display onto second lens 324 , either by itself or at the same time that first image generator 360 is displaying augmented reality display 370 onto first lens 322 .
- both first image generator and second image generator are deactivated.
- one or both image generators are configured to display augmented reality displays on an external surface or medium distinct from the lenses.
- visual training aid 310 is supported by eyeglasses 321 .
- frame 327 supports processor 330 , memory 340 , global positioning system 350 , and clock 390 in compartment 325 .
- Frame 327 further supports first and second image generators 360 and 362 and first and second lenses 322 and 324 .
- visual training aid 310 may be described as a self-contained or standalone embodiment.
- Visual training aid 410 includes many similar or identical features to visual training aids 210 and 310 combined in unique and distinct ways. Thus, for the sake of brevity, each feature of visual training aid 410 will not be redundantly explained. Rather, key distinctions between visual training aids 210 and 310 and visual training aid 410 will be described in detail and the reader should reference the discussion above for features substantially similar between the different visual training aids.
- visual training aid 410 includes components separated from one another and connected by data wires or cables 436 and 437 . Whereas visual training aid 310 was self-contained, visual training aid 410 includes a system of components working together to achieve visual training aid functions substantially similar to those described above. The reader can further see in FIG. 5 that visual training aid 410 includes a camera 480 and an external device 434 in addition to an eyewear article 420 in the form of eyeglasses 421 with a lens 422 , a processor 430 , memory 440 , a global positioning system 450 , a image generator 460 , and a clock 490 .
- Data cables 436 and 437 may be any conventional data cable suitable for exchanging applicable data between devices.
- a variety of analog and digital data transfer technologies and interfaces may be employed, including interfaces such as the universal serial bus (USB) interface, the high-definition multimedia interface, (HDMI), the digital video interface (DVI), the IEEE 1394 interface, and the like.
- USB universal serial bus
- HDMI high-definition multimedia interface
- DVI digital video interface
- IEEE 1394 IEEE 1394 interface
- External device 434 houses processor 430 , memory 440 , and clock 490 .
- the external device houses a global positioning system as well.
- external device 434 includes a display screen 435 , which is configured to display information to a user.
- the display screen is a touch screen configured to read touch inputs by a user.
- the external device may be a wide variety of commercially available electronic devices designed for functions beyond those applicable to a visual training aid.
- the external device may be an Apple iPhone® brand mobile phone, mobile phones in general, an Apple iTouch® brand mobile device, a mobile device designed with the Android® operating system, portable media players, personal computers, personal data assistants, watches, and other mobile computing devices that are capable of executing third-party developed software.
- External device 434 is programmed with software instructions and drivers necessary for processor 430 to achieve the functions described herein of a visual training aid.
- Camera 480 is mounted to eyeglasses 421 in a position to capture course image data.
- camera 480 is configured to dynamically target a portion of the course and capture course image data corresponding to where a user wearing the eyeglasses is looking.
- Processor 430 is configured to dynamically read the course image data from camera 480 and store it in memory 440 , in addition to other functions, such as correlating the course image data with time interval data from clock 490 and position data from global positioning system 450 .
- the visual training aid includes a display monitor
- the course image data is displayed on the display monitor to assist the user to perceive the course.
- Visual training aid 510 includes many similar or identical features to visual training aids 210 , 310 , and 410 combined in unique and distinct ways. Thus, for the sake of brevity, each feature of visual training aid 510 will not be redundantly explained. Rather, key distinctions between visual training aids 210 , 310 , and 410 and visual training aid 510 will be described in detail and the reader should reference the discussion above for features substantially similar between the different visual training aids.
- visual training aid 510 includes components separated from one another and in data communication with one another with a first wireless communications link 536 and a second wireless communications link 537 .
- visual training aid 310 was self-contained and the components of visual training aid 410 were in data communication with data cables 436 and 437
- visual training aid 510 includes a system of components working together wirelessly to achieve visual training aid functions substantially similar to those described above. The reader can further see in FIG.
- visual training aid 510 includes a front camera 580 and a rear camera 582 in addition to an eyewear article 520 in the form of eyeglasses 521 with a frame 527 and a lens support 528 , a display monitor 526 , an external device 534 , a processor 530 , memory 540 , a global positioning system 550 , and a clock 590 .
- Wireless data communications links 536 and 537 may be any conventional wireless data communications link for exchanging applicable data between devices.
- the wireless data communications links may employ communications standards such as WiFiTM, BluetoothTM, WiMax, and the like. Any conventional or latter developed wireless data communications link technology may be used with visual training aid 510 .
- Front camera 580 targets a portion of the course disposed in front of display monitor 526 .
- Rear camera 582 targets a portion of the course behind the user.
- front camera 580 is positioned to capture forward course image data corresponding to images of the course forward of a user wearing eyeglasses 521 and rear camera 582 is positioned to capture rearward course image data corresponding to images of the course behind the user.
- Front camera 580 and rear camera 582 may be the same or different and may each be configured to capture still images or a sequence of still images to define a video.
- Display monitor 526 is mounted to eyeglasses 521 in lens support 528 .
- display monitor 526 is at least partially transparent to enable a user wearing eyeglasses 521 to see the course through display monitor 526 .
- course image data collected from one or more of front camera 580 and rear camera 582 and displayed on display monitor 526 via instructions from processor 530 further assists the user to perceive the course.
- the course image data collected by the front and rear cameras and displayed on the display monitor via instructions from the processor may be the primary means for the user to perceive the course by at least one of his eyes.
- the virtual training aids described herein may be configured to connect to the internet, although this functionality is not required.
- This disclosure contemplates the use of internet connectivity by internet capable visual training aids to store and retrieve training data between computers and devices connected via the internet. Further, two or more visual training aids may share training data between them, in real-time or over time spans, via the internet. Uploading training data to social networking sites and applications is also contemplated.
- Visual training aids may additionally include an attached or integrated audio device.
- the audio devices may be configured to output audio signals to headphones, ear buds, or other portable means of playing audio. Audio devices may be useful in providing additional training data or for playing stored audio during a workout.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Computer Hardware Design (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Optics & Photonics (AREA)
- Physical Education & Sports Medicine (AREA)
- Environmental & Geological Engineering (AREA)
- Epidemiology (AREA)
- Public Health (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Atmospheric Sciences (AREA)
- Toxicology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Ophthalmology & Optometry (AREA)
- Multimedia (AREA)
- Otolaryngology (AREA)
- General Engineering & Computer Science (AREA)
- Rehabilitation Tools (AREA)
Abstract
Description
- This application claims priority to copending U.S. application Ser. No. 12/869,481, filed on Aug. 26, 2010, which is an application claiming the benefit of U.S. Provisional Application Ser. No. 61/329,648, filed on Apr. 30, 2010, under section 35 USC 119(c). Both of U.S. application Ser. No. 12/869,481 and U.S. Provisional Application Ser. No. 61/329,648 are hereby incorporated by reference for all purposes.
- The present disclosure relates generally to visual training devices, systems, and methods, collectively visual training aids. In particular, location aware visual training aids are described herein.
- Training assistance devices in the art are not entirely satisfactory. Specifically, training aids in the art provide users with limited training data during a workout. Training data typically provided by conventional training devices includes current and average speed information, pace information, route information, the time in which the user completes a particular route, current and average heart rate information, and current or average cadence information.
- Training assistance devices in the art that provide visual data often require users to look away from their path to a separate device, such as a watch or mobile device. Both looking away from the path and handling other devices can have detrimental effect on an athlete's pace and route. Additionally, these distractions may cause safety concerns due to the lack of focus on the route.
- The advancement of video eyewear technologies now allows users to view computer generated graphics on a display proximate their eyes and substantially cover their entire field of vision. Such technologies generally achieve this functionality by attaching projectors to a set of eyewear that are capable of projecting graphical images on the lenses. These displays may receive a computer generated image from an external source and use a series of mirrors to project this image on the lens, thereby placing computer generated graphical images in the user's natural line of vision. There are, however, other technologies by which images are displayed on the lenses of such video eyewear, including more traditional display technologies such as LCD.
- These video eyewear technologies are particularly suited to the display of “augmented reality” displays. “Augmented reality” displays comprise a computer generated graphical display laid over a portion of a user's or mobile device's natural field of vision. These “augmented reality” displays allow a user to view computer generated images including data related to objects in their natural field of vision. Augmented reality displays may include any display including both natural and computer generated elements.
- Additionally, the advancement of mobile computing devices have allowed for the development of substantially featured computing devices that afford great portability. These devices often include software development kits and application programming interfaces that allow for the development of software and drivers to allow the driver to interact with a wide array of external hardware. Additionally, these devices often allow connection to the internet and local networks through various protocols. These devices may also include functionality that acquire data relating to location and movement, including, but not limited to, global positioning system receivers, accelerometers, magnetometers, and pedometers.
- Examples of mobile computing devices include Apple iPhone mobile devices, mobile devices designed with the Android operating system, other mobile phone devices, portable media players, personal computers, watches, and other mobile computing devices that are capable of executing third party developed software.
- The present disclosure is directed to visual training aids including an eyewear article including a lens, an image generator mounted to the eyewear article in a position to display an image on the lens, a processor operatively connected to and in data communication with the image generator, a global positioning system operatively connected to and in data communication with the processor, a computer readable medium operatively connected to and in data communication with the processor. In some examples, the visual training aid includes a pair of eyeglasses. In some examples, the visual training aid includes a display monitor mounted to the eyewear. In some examples, the visual training aid includes a camera mounted to the eyewear.
-
FIG. 1 is a schematic view of a computing device that may be used to implement various aspects of visual training aids. -
FIG. 2 is a schematic view of a first example of a visual training aid. -
FIG. 3 is a rear elevation view of a visual training aid including a pair of eyeglasses. -
FIG. 4 is a side elevation view of the visual training aid shown inFIG. 3 . -
FIG. 5 illustrates an example of a visual training aid with a forward facing camera. -
FIG. 6 illustrates an example of a visual training aid with a forward facing camera and a rearward facing camera and where components of the visual training aid are connected by a wireless data communications link. - The disclosed visual training aids will become better understood through review of the following detailed description in conjunction with the figures. The detailed description and figures provide merely examples of the various inventions described herein. Those skilled in the art will understand that the disclosed examples may be varied, modified, and altered without departing from the scope of the inventions described herein. Many variations are contemplated for different applications and design considerations; however, for the sake of brevity, each and every contemplated variation is not individually described in the following detailed description.
- Throughout the following detailed description, examples of various visual training aids are provided. Related features in the examples may be identical, similar, or dissimilar in different examples. For the sake of brevity, related features will not be redundantly explained in each example. Instead, the use of related feature names will cue the reader that the feature with a related feature name may be similar to the related feature in an example explained previously. Features specific to a given example will be described in that particular example. The reader should understand that a given feature need not be the same or similar to the specific portrayal of a related feature in any given figure or example.
- Various examples of visual training aids may be implemented using electronic circuitry configured to perform one or more functions. For example, with some embodiments of the invention, the visual training aid may be implemented using one or more application-specific integrated circuits (ASICs). More typically, however, components of various examples of the invention will be implemented using a programmable computing device executing firmware or software instructions, or by some combination of purpose-specific electronic circuitry and firmware or software instructions executing on a programmable computing device.
- Accordingly,
FIG. 1 shows one illustrative example of acomputer 101 that can be used to implement various embodiments of the invention.Computer 101 may be incorporated within a variety of consumer electronic devices, such as personal media players, cellular phones, smart phones, personal data assistants, global positioning system devices, and the like. - As seen in this figure,
computer 101 has acomputing unit 103.Computing unit 103 typically includes aprocessing unit 105 and asystem memory 107.Processing unit 105 may be any type of processing device for executing software instructions, but will conventionally be a microprocessor device.System memory 107 may include both a read-only memory (ROM) 109 and a random access memory (RAM) 111. As will be appreciated by those of ordinary skill in the art, both read-only memory (ROM) 109 and random access memory (RAM) 111 may store software instructions to be executed byprocessing unit 105. -
Processing unit 105 andsystem memory 107 are connected, either directly or indirectly, through abus 113 or alternate communication structure to one or more peripheral devices. For example,processing unit 105 orsystem memory 107 may be directly or indirectly connected to additional memory storage, such as ahard disk drive 115, a removable magnetic disk drive 117, anoptical disk drive 119, and aflash memory card 121.Processing unit 105 andsystem memory 107 also may be directly or indirectly connected to one ormore input devices 123 and one or more output devices 125.Input devices 123 may include, for example, a keyboard, touch screen, a remote control pad, a pointing device (such as a mouse, touchpad, stylus, trackball, or joystick), a scanner, a camera or a microphone. Output devices 125 may include, for example, a monitor display, an integrated display, television, printer, stereo, or speakers. - Still further, computing
unit 103 will be directly or indirectly connected to one ormore network interfaces 127 for communicating with a network. This type ofnetwork interface 127, also sometimes referred to as a network adapter or network interface card (NIC), translates data and control signals fromcomputing unit 103 into network messages according to one or more communication protocols, such as the Transmission Control Protocol (TCP), the Internet Protocol (IP), and the User Datagram Protocol (UDP). These protocols are well known in the art, and thus will not be discussed here in more detail. Aninterface 127 may employ any suitable connection agent for connecting to a network, including, for example, a wireless transceiver, a power line adapter, a modem, or an Ethernet connection. - It should be appreciated that, in addition to the input, output and storage peripheral devices specifically listed above, the computing device may be connected to a variety of other peripheral devices, including some that may perform input, output and storage functions, or some combination thereof. For example, the
computer 101 may be connected to a digital music player, such as an IPOD® brand digital music player or IPHONE® brand smartphone available from Apple, Inc. of Cupertino, Calif. As known in the art, this type of digital music player can serve as both an output device for a computer (e.g., outputting music from a sound file or pictures from an image file) and a storage device. In addition, this type of digital music play also can serve as an output device for outputting visual training information, as will be discussed in more detail below. - In addition to a digital music player,
computer 101 may be connected to or otherwise include one or more other peripheral devices, such as a telephone. The telephone may be, for example, a wireless “smart phone.” As known in the art, this type of telephone communicates through a wireless network using radio frequency transmissions. In addition to simple communication functionality, a “smart phone” may also provide a user with one or more data management functions, such as sending, receiving and viewing electronic messages (e.g., electronic mail messages, SMS text messages, etc.), recording or playing back sound files, recording or playing back image files (e.g., still picture or moving video image files), viewing and editing files with text (e.g., Microsoft Word or Excel files, or Adobe Acrobat files), etc. Because of the data management capability of this type of telephone, a user may connect the telephone withcomputer 101 so that their data maintained may be synchronized. - Of course, still other peripheral devices may be included with our otherwise connected to a
computer 101 of the type illustrated inFIG. 1 , as is well known in the art. In some cases, a peripheral device may be permanently or semi-permanently connected tocomputing unit 103. For example, with many computers,computing unit 103, hard disk drive 117, removableoptical disk drive 119 and a display are semi-permanently encased in a single housing. - Still other peripheral devices may be removably connected to
computer 101, however.Computer 101 may include, for example, one or more communication ports through which a peripheral device can be connected to computing unit 103 (either directly or indirectly through bus 113). These communication ports may thus include a parallel bus port or a serial bus port, such as a serial bus port using the Universal Serial Bus (USB) standard or the IEEE 1394 High Speed Serial Bus standard (e.g., a Firewire port). Alternately or additionally,computer 101 may include a wireless data “port,” such as a Bluetooth interface, a Wi-Fi interface, an infrared data port, or the like. - It should be appreciated that a computing device employed according various examples of the invention may include more components than
computer 101 illustrated inFIG. 1 , fewer components thancomputer 101, or a different combination of components thancomputer 101. Some implementations of the invention, for example, may employ one or more computing devices that are intended to have a very specific functionality, such as a digital music player or server computer. These computing devices may thus omit unnecessary peripherals, such as thenetwork interface 115, removableoptical disk drive 119, printers, scanners, external hard drives, etc. Some implementations of the invention may alternately or additionally employ computing devices that are intended to be capable of a wide variety of functions, such as a desktop or laptop personal computer. These computing devices may have any combination of peripheral devices or additional components as desired. - With reference to
FIG. 2 , avisual training aid 210 includeseyeglasses 220, aprocessor 230,memory 240, aglobal positioning system 250, and animage generator 260.Visual training aid 210 further includes an optional image capturing device or camera or 280 to capture images of the terrain surrounding the user. Additional components are included in certain examples, such as laser range finders and radar systems. Appropriate clocks and other time keeping devices, whether physical or software based, are contemplated for measuring relevant time intervals. Additionally or alternatively, the visual training aid may include data cables, wireless data transmitters and receivers, and removable computer readable storage mediums. - As will be described in more detail below, visual training aids, such as
visual training aid 210, provide users with dynamically updated visual training data during a workout or athletic activity. Visual training aids described herein may be used in a variety of activities, including running, skiing, snowboarding, bicycling, horseback riding, racecar driving, motorcycle driving, and other timed competitive endeavors. The visual data helps to provide the user with an awareness of his position relative to the position of a competitor or to his own prior performance on a course. This visual data may thus stimulate the user's competitive drive and enhance his workout performance and training satisfaction. - Indeed, the visual training aids enhance training effectiveness and motivation through simulating elements of competition during solo training endeavors. With brief reference to a
visual training aid 310 shown inFIG. 3 for demonstrating an example of simulated competition provided by the visual training aids described herein, the reader can see that a user utilizingvisual training aid 310 sees anaugmented reality display 370 including a computer generated representation of afirst person 372 and a computer generated image of asecond person 374. A person in this context may also be referred to as a “competitor,” a “virtual competitor,” or a “prior performer.” - The computer generated images of the first and second persons may be either static or animated and may depict another person's performance on the course or the user's own past performance on the course. The computer generated images depicting the apparent position of the
372 and 374 relative to the wearer and the surroundings is continually updated based on the position ofvirtual competitors 372 and 374 on the course during a prior training session relative the user's current position at the same given time interval.competitors - For example, if the user's performance falls behind the competitor's prior pace on the course, representations of the competitor are displayed such that they appear in front of the user from the user's current vantage point. If the user's performance is faster than the competitor's prior pace, representations of the competitor are displayed such that they appear behind the user from the user's vantage point. If a user's pace is substantially the same as the competitor's prior pace on the course, representations of the competitor are projected to appear to the side of the user. The relative position of the trainer and representation of the prior performer will be constantly changing as their respective velocities change throughout the course.
- By viewing these displays of a virtual competitor, the user will see how his performance is comparing to a prior performance or a desired hypothetical performance. Viewing displays of dynamically updated virtual competitors also allow the user to simulate competition, making for a more enjoyable and effective training experience.
- With reference to
FIG. 2 , the components ofvisual training aid 210 will now be described.Eyewear 220 serves to support selected components ofvisual training aid 210 and to position an augmented reality display in a position to be viewed by a user of the visual training aid. Eyewear may include monocles, goggles, head mounted displays, helmet mounted displays, or any other arrangement of lenses supported proximate the eye. - The reader should understand that the visual training aids described herein do not require eyewear. Indeed, in some examples the computer generated display of virtual competitors is displayed on a mobile computing device, such as the display or screen of a mobile phone, including an iPhone® brand mobile phone, a personal media player, including the iTouch® personal media player, a personal data assistant, and the like.
- In examples without eyewear, the user carries the mobile device with him during a training session and looks at the screen, for example, by holding the mobile device in front of him, to see on the screen or display the position of virtual competitors relative to his own position. In one example, the augmented reality display on the mobile device includes a map depicting the course, similar to a global positioning system map depicting a route, with virtual competitors positioned relative to the user's current position on the map. In other examples, the augmented reality display is a real-time image of the surroundings visible through a camera lens of the mobile device with computer generated virtual competitors represented on the display in a position reflecting the difference in position between the user's current position and the person's position on the course during a prior training session at the same given time interval.
-
Processor 230 is programmed with instructions for receiving data from components ofvisual training aid 210, includingmemory 240,global positioning system 250, camera 280, and clock ortimer 290. Data from components ofvisual training aid 210, such asglobal positioning system 250, camera 280, andclock 290 may be in a raw format that is not directly usable by other components ofvisual training aid 210, such asimage generator 260 ormemory 240. Accordingly, processor is programmed with instructions for translating, computing, converting, correlating, and integrating (collectively “processing”) data from each component into usable formats. - For example, processor is programmed with instruction for processing prior position data and prior time interval data from
memory 240, current position data fromglobal positioning system 250, terrain image data from camera 280, and current time interval data fromclock 290. The processed data may be stored intomemory 240, communicated to databases or devices over the internet, or sent to other components ofvisual training aid 210, such asimage generator 260. Althoughprocessor 230 may process and store processed data,visual training aid 210 storing raw data intomemory 240 or into other databases is equally within this disclosure. Data collected by components ofvisual training aid 210 may be displayed in a readable way, stored and retrieved to local or remote storage means, including mobile devices, cloud technologies, internet servers, and social networking applications and sites. - Reading past position data from
memory 240 may include reading multiple sets of past position data frommemory 240. For example, each set of past position data may correspond to a person's position on the course during prior training sessions at a given time interval relative to the respective course start times for each prior training session. The past position data may correspond to different people's position or a single person's position during two or more prior training sessions. -
Processor 230 is programmed with instruction for determining a current position of the user or eyewear article on a course at a given time interval relative to a current course start time. To make these determinations,processor 230 dynamically current position data fromglobal positioning system 260 and time interval data fromclock 290. Additionally or alternatively, the processor may receive one or more of these data from memory. -
Processor 230 is further programmed with instructions for generating competitor image data depicting a person's position on a course relative to the user's current position at a given time interval. The competitor image data may include data corresponding to images of multiple competitors. The competitor image data includes instructions forimage generator 260 to display an image on the lens depicting a person's or competitor's position on the course during a prior training session relative to the current position of the user or ofeyewear article 220 at a given time interval. - The competitor image data may include various parameters determining how the image of the virtual competitor will be displayed. For example, the competitor image data may include where in the user's field of view the computer generated image of the virtual competitor is to be displayed to account for where the user is currently looking. Additionally or alternatively, the competitor image data may include the size of the image of the virtual competitor to display to account for proper distance and perspective. Further, the competitor image data may include the profile of the virtual competitor to display, e.g., a view of the virtual competitor's back, left side, right side, or front, to account for where the user is looking and the respective position of the user and the person's prior position on the course at the relevant time interval.
- For example,
processor 230 may be programmed with instructions to dynamically adjust the direction in which the graphical representation of a prior performer's position is displayed based on dynamically updated data representing the direction the user is currently looking. For example, if the user is looking straight ahead and the competitor's prior performance at that given time interval was ahead of the user's current position, the processor instructions specify that the graphical representation of the prior performer be displayed to appear in front of the user. If the user were to look to his right 60°, the processor instructions specify that the graphical representation of the prior performer be displayed to appear 600 to the left of the user. - The visual training aids described herein may include any conventional means for detecting the direction a user is looking. For example, the visual training aid may include eye tracking devices, accelerometers to detect head position and orientation, image capturing devices to capture images of the terrain around the wearer, laser range finders, radar, and computer vision technologies.
-
Processor 230 is programmed with instructions for outputting or sending data to components ofvisual training aid 210, includingmemory 240 andimage generator 260. For example,processor 230 is programmed with instructions for sending current position data and current time interval data tomemory 250. Additionally,processor 230 is programmed with instructions for sending competitor image data to imagegenerator 260, including competitor image data for multiple virtual competitors. -
Processor 230 may be programmed with instructions to generate competitor image data including an animated depiction of a person on the course during a prior training session relative to the current position of the user or eyewear article at the given time interval. An animated depiction of the person or virtual competitor has been observed to increase the realism of the virtual competitor and to enhance the competitive effect of “racing” against the virtual competitor. -
Processor 230 may be any conventional processor suitable for inputting, processing, and outputting data consistent with the functions and programmed instructions described herein. The processor may include multiple processors programmed to carry out specific instructions in concert. The processor may be specially fabricated for use in the visual training aids described herein or may employ commercially available processors or processors used for other devices. - For example, the processor used by visual training aids may be a processor incorporated within an Apple iPhone brand mobile phone, which is designed for functions beyond those applicable to a visual training aid. User installed software may provide all necessary instructions and drivers to achieve the functions described herein of a visual training aid. Other suitable processors include those utilized by Apple iTouch® brand mobile devices, mobile devices designed with the Android® operating system, other mobile devices, including portable media players, personal computers, watches, and other mobile computing devices that are capable of executing third-party developed software.
- Additionally or alternatively, proprietary processor systems designed specifically for use with visual training aids describe herein are equally within this disclosure. In such designs, the processor system may include operating system software.
-
Memory 240 may comprise any conventional data storage medium known in the art. For example,memory 240 may include magnetic media, optical media, flash media, and the like.Memory 240 may also include a plurality of memory units, each memory unit being the same media or one or more being different types of media. - In the example shown in
FIG. 2 ,global positioning system 250 dynamically determines the current, real-time position of the user as he moves about a course. This data may be described as current position data. Additionally or alternatively, current position data may be obtained from external sensors that monitor the position of the user and/or a component of the visual training aid. Other technologies, including pedometers, accelerometers, and magnetometers, may work in conjunction with global positioning systems, and in some examples replace them, to generate position and other training data. - Any conventional global positioning system may be used for generating current position data. For example, commercially available, stand-alone global positioning systems as well as global positioning systems incorporated into other devices, such as cellular phones and portable media players, are suitable. Appropriate software, cables, or wireless data protocols may be employed to communicate the current position data from the global positioning system to the processor. In some examples, the processor used by a global positioning system itself may be programmed to carry out the functions and instructions of the visual training aids described herein.
- The current position data is sent to
processor 230 for generating instructions for use byimage generator 260. Specifically,processor 230 uses the current position data and time stamped past position data of a prior person (or persons) stored inmemory 240 to determine the position of the user relative to the person's position on the course during a prior training session at a given time interval. Time stamped position data refers to position data cross referenced with time interval data detailing how long into a training session person was at a given position.Processor 230 then sends image data to imagegenerator 260 for displaying an image to the user depicting the person's position on the course during the prior training session relative to the current position of the user at the given time interval. -
Processor 230 may be configured to use the current position data in conjunction with timing data fromclock 290 to generate a wide range of training data. The training data may include the user's current speed, average speed, cadence information, and workout duration among various other useful metrics relating to the current workout, including previous pace in which the same route was completed. The training data may also include cues for the user, including graphical cues assisting the user to follow a route, such as maps and arrows, and cues assisting the user in knowing when to begin a particular workout. The training data may be displayed to the user over the course of a workout and/or may be stored inmemory 240 for later retrieval. - In addition to the training data described above,
processor 230 is configured to use the current position data in conjunction with timing data fromclock 290 to generate and store a set of time-stamped position data for use in subsequent training sessions. The time stamped position data generated during a current training session can serve as past position data during subsequent training sessions, either by a 3rd party trainee or by the user himself. That is, the user's time stamped position data processed byprocessor 230 and stored inmemory 240 may be used byprocessor 230 during a subsequent training session as the basis for generating a virtual competitor for the user of visual training aid. -
Image generator 260 displays an image on the lens for viewing by the user ofvisual training aid 210 based on competitor image data supplied byprocessor 230. In some examples, the image generator is configured to display a series of images in succession to define a video or an animated depiction of the virtual competitor. -
Image generator 260 may include a wide variety of image display technologies. In some examples, the image generator includes a projector configured to project an image or a series of images onto the lens. In other examples, the image generator includes a hologram generating system for generating holograms viewable by the user. In some examples, the image generator includes a heads-up-display projection system. - In still further other examples, the image generator is configured to display in image visible to the user on a screen or display that at least partially blocks a user's view through the lens. For example, the visual training aid may include an opaque screen positioned across at least a portion of at least one of the lenses in a pair of eyeglasses and a projector configured to project an image onto the screen. In another example, the image generator includes a display monitor, such as a liquid crystal display, positioned across at least a portion of a lens for displaying an image or a series of images to a user. While the user is unable to see through the screen or the display monitor, the image generator may display images corresponding to the user's surroundings based on terrain image data supplied by 230 processor in conjunction with camera 280.
- Camera 280 serves to capture images of a course, including the user's surroundings and the corresponding terrain. Data corresponding to images of the course and surroundings may be described as terrain image data or course image data. When displayed on the lens of
eyewear 220, the terrain image data enables the user to perceive his surroundings even if he does not have a direct line of sight to his surroundings. For example, the terrain image data may enable the user to perceive his surroundings when an opaque screen or display monitor is positioned in front of one of more of his eyes. - In some examples, camera 280 is a video camera configured to capture a series of images in succession to define a video. A wide variety of acceptable cameras and/or video cameras are commercially available. The resolution of the cameras may be selected to optimize data throughput and image quality. Acceptable resolutions include interlaced and progressive image scans at 480, 525, 720, 1080 vertical scanning lines, as well as vertical scanning lines in between, higher, and lower than these example values.
- In some examples, the visual training aid includes two or more cameras. In one example, a first camera is positioned to capture forward course image data corresponding to images of the course forward of a user wearing the eyewear and a second camera is positioned to capture rearward course image data corresponding to images of the course behind the user. In other examples, a sufficient number of cameras are used to capture images providing a substantially 360° view of the course.
- The reader should understand that the camera is an optional feature of the visual training aids described herein. In some examples, the visual training aid does not include a camera. For instance, when the image generator is configured to display images of virtual competitors on a substantially transparent lens, the user can see the course and the terrain through the lens without the need for a camera. In these examples, the depictions of the virtual competitors are integrated with the actual view of the course and the terrain through the lens by where and how they are displayed on the lens.
-
Clock 290 may be any conventional mechanical or software based time keeping device. In addition to a host of conventional functions a clock may serve in conjunction with a processor in a computing device,clock 290 serves as a timer for generating time interval data. - The time interval data corresponds to the time difference between a given time on a course after starting a training session on the course. A current time interval is applicable to a user's current training session and is defined as the user's current time on a course minus the user's current course start time. A past time interval is applicable to a person's prior training session on a course and is defined as the time difference between a given time on the course and the person's past course start time.
- The current time interval will dynamically increase with the passage of time after a current course start time.
Clock 290 is configured to dynamically output time interval data toprocessor 230.Processor 230 uses the current time interval data to determine which past position data is applicable for generating image data to imagegenerator 260. In particular,processor 230 will select past position data having a past time interval equal to the current time interval to correlate a person's position on the course during a prior training session with a user's current position on a course at a same given time interval. In this manner, the relative position of the virtual competitor at a given time interval corresponding to how long the user has currently been on the course can be displayed to the user. - Turning attention to
FIGS. 3 and 4 , avisual training aid 310 will now be described.Visual training aid 310 includes many similar or identical features tovisual training aid 210 combined in unique and distinct ways. Thus, for the sake of brevity, each feature ofvisual training aid 310 will not be redundantly explained. Rather, key distinctions betweenvisual training aid 210 andvisual training aid 310 will be described in detail and the reader should reference the discussion above for features substantially similar between the two visual training aids. - With reference to
FIGS. 3 and 4 ,visual training aid 310 includes aneyewear article 320 in the form ofeyeglasses 321, aprocessor 330,memory 340, aglobal positioning system 350, afirst image generator 360, asecond image generator 362, and a clock 390. As shown inFIG. 3 ,visual training aid 310 generates anaugmented reality display 370 to serve as a training aid for a user. As shown inFIG. 4 , all components ofvisual training aid 310 is supported byeyewear 320. -
Eyeglasses 321 supports components ofvisual training aid 310 and positions augmentedreality display 370 in a position to be viewed by a user. In the example shown inFIG. 3 ,eyeglasses 321 include afirst lens 322, asecond lens 324, and a frame 338. In other examples, the eyewear article includes a single lens. - First and
322 and 324 are made of a substrate that is generally suitable for the display of an image on their surface for transmission of light through the lens. In some examples, one or both of the lenses are configured to transmit light primarily in a first direction towards the user's eye, with reduced or no transmission of light in the opposite direction. A wide range of conventional lenses may be used.second lenses -
Frame 327 supports afirst lens support 328 and asecond lens support 329.Frame 327 may be made from any material conventionally used to construct eyeglass frames, including plastics and metals. The frame may adopt a wide variety of conventional eyewear frame designs. However, in the example depicted inFIG. 4 ,frame 327 is specially designed with acompartment 325 for housing components ofvisual training aid 310, includingprocessor 330,memory 340,global positioning system 350, and clock 390. Additional components, such as batteries, may be housed incompartment 325 as well. -
Visual training aid 310 can be operated in multiple modes. For example, modes of operation include a mode where bothfirst image generator 360 andsecond image generator 362 are displaying augmented reality displays, a mode where just one of the first and second image generators are displaying an augmented reality display, and a mode where neither the first and second image generators are displaying an augmented reality display. The mode of operation depicted inFIG. 3 is a mode where onlyfirst image generator 360 is displaying anaugmented reality display 370 on onlyfirst lens 322. - However,
visual training aid 310 may be operated in a mode wheresecond image generator 362 is displaying an augmented reality display ontosecond lens 324, either by itself or at the same time thatfirst image generator 360 is displayingaugmented reality display 370 ontofirst lens 322. In another mode, both first image generator and second image generator are deactivated. In some examples, one or both image generators are configured to display augmented reality displays on an external surface or medium distinct from the lenses. - As can be seen in
FIG. 4 , all of the components ofvisual training aid 310 are supported byeyeglasses 321. In particular,frame 327 supportsprocessor 330,memory 340,global positioning system 350, and clock 390 incompartment 325.Frame 327 further supports first and 360 and 362 and first andsecond image generators 322 and 324. Thus,second lenses visual training aid 310 may be described as a self-contained or standalone embodiment. - Turning attention to
FIG. 5 , avisual training aid 410 will now be described.Visual training aid 410 includes many similar or identical features to visual training aids 210 and 310 combined in unique and distinct ways. Thus, for the sake of brevity, each feature ofvisual training aid 410 will not be redundantly explained. Rather, key distinctions between visual training aids 210 and 310 andvisual training aid 410 will be described in detail and the reader should reference the discussion above for features substantially similar between the different visual training aids. - With reference to
FIG. 5 , the reader can see thatvisual training aid 410 includes components separated from one another and connected by data wires or 436 and 437. Whereascables visual training aid 310 was self-contained,visual training aid 410 includes a system of components working together to achieve visual training aid functions substantially similar to those described above. The reader can further see inFIG. 5 thatvisual training aid 410 includes acamera 480 and anexternal device 434 in addition to aneyewear article 420 in the form ofeyeglasses 421 with alens 422, aprocessor 430,memory 440, aglobal positioning system 450, aimage generator 460, and aclock 490. -
436 and 437 may be any conventional data cable suitable for exchanging applicable data between devices. A variety of analog and digital data transfer technologies and interfaces may be employed, including interfaces such as the universal serial bus (USB) interface, the high-definition multimedia interface, (HDMI), the digital video interface (DVI), the IEEE 1394 interface, and the like. As can be seen inData cables FIG. 5 ,data cable 436 connectsglobal positioning system 450 withprocessor 430 housed inexternal device 434.Data cable 437 connectsexternal device 434 withimage generator 460 supported byeyeglasses 421. -
External device 434houses processor 430,memory 440, andclock 490. In some examples, the external device houses a global positioning system as well. As can be seen inFIG. 5 ,external device 434 includes adisplay screen 435, which is configured to display information to a user. In some examples, the display screen is a touch screen configured to read touch inputs by a user. - The external device may be a wide variety of commercially available electronic devices designed for functions beyond those applicable to a visual training aid. For example, the external device may be an Apple iPhone® brand mobile phone, mobile phones in general, an Apple iTouch® brand mobile device, a mobile device designed with the Android® operating system, portable media players, personal computers, personal data assistants, watches, and other mobile computing devices that are capable of executing third-party developed software.
External device 434 is programmed with software instructions and drivers necessary forprocessor 430 to achieve the functions described herein of a visual training aid. -
Camera 480 is mounted toeyeglasses 421 in a position to capture course image data. In the example shown inFIG. 5 ,camera 480 is configured to dynamically target a portion of the course and capture course image data corresponding to where a user wearing the eyeglasses is looking. -
Processor 430 is configured to dynamically read the course image data fromcamera 480 and store it inmemory 440, in addition to other functions, such as correlating the course image data with time interval data fromclock 490 and position data fromglobal positioning system 450. In examples where the visual training aid includes a display monitor, the course image data is displayed on the display monitor to assist the user to perceive the course. - Turning attention to
FIG. 6 , avisual training aid 510 will now be described.Visual training aid 510 includes many similar or identical features to visual training aids 210, 310, and 410 combined in unique and distinct ways. Thus, for the sake of brevity, each feature ofvisual training aid 510 will not be redundantly explained. Rather, key distinctions between visual training aids 210, 310, and 410 andvisual training aid 510 will be described in detail and the reader should reference the discussion above for features substantially similar between the different visual training aids. - With reference to
FIG. 6 , the reader can see thatvisual training aid 510 includes components separated from one another and in data communication with one another with a first wireless communications link 536 and a second wireless communications link 537. Whereasvisual training aid 310 was self-contained and the components ofvisual training aid 410 were in data communication with 436 and 437,data cables visual training aid 510 includes a system of components working together wirelessly to achieve visual training aid functions substantially similar to those described above. The reader can further see inFIG. 6 thatvisual training aid 510 includes afront camera 580 and arear camera 582 in addition to aneyewear article 520 in the form ofeyeglasses 521 with aframe 527 and alens support 528, adisplay monitor 526, anexternal device 534, aprocessor 530, memory 540, aglobal positioning system 550, and aclock 590. - Wireless
536 and 537 may be any conventional wireless data communications link for exchanging applicable data between devices. For example, the wireless data communications links may employ communications standards such as WiFi™, Bluetooth™, WiMax, and the like. Any conventional or latter developed wireless data communications link technology may be used withdata communications links visual training aid 510. -
Front camera 580 targets a portion of the course disposed in front ofdisplay monitor 526.Rear camera 582 targets a portion of the course behind the user. Expressed another way,front camera 580 is positioned to capture forward course image data corresponding to images of the course forward of auser wearing eyeglasses 521 andrear camera 582 is positioned to capture rearward course image data corresponding to images of the course behind the user.Front camera 580 andrear camera 582 may be the same or different and may each be configured to capture still images or a sequence of still images to define a video. -
Display monitor 526 is mounted toeyeglasses 521 inlens support 528. In the particular example shown inFIG. 6 , display monitor 526 is at least partially transparent to enable auser wearing eyeglasses 521 to see the course throughdisplay monitor 526. However, course image data collected from one or more offront camera 580 andrear camera 582 and displayed on display monitor 526 via instructions fromprocessor 530 further assists the user to perceive the course. In examples where the display monitor is not at least partially transparent, the course image data collected by the front and rear cameras and displayed on the display monitor via instructions from the processor may be the primary means for the user to perceive the course by at least one of his eyes. - The virtual training aids described herein may be configured to connect to the internet, although this functionality is not required. This disclosure contemplates the use of internet connectivity by internet capable visual training aids to store and retrieve training data between computers and devices connected via the internet. Further, two or more visual training aids may share training data between them, in real-time or over time spans, via the internet. Uploading training data to social networking sites and applications is also contemplated.
- Visual training aids may additionally include an attached or integrated audio device. The audio devices may be configured to output audio signals to headphones, ear buds, or other portable means of playing audio. Audio devices may be useful in providing additional training data or for playing stored audio during a workout.
- The disclosure above encompasses multiple distinct inventions with independent utility. While each of these inventions has been disclosed in a particular form, the specific embodiments disclosed and illustrated above are not to be considered in a limiting sense as numerous variations are possible. The subject matter of the inventions includes all novel and non-obvious combinations and subcombinations of the various elements, features, functions and/or properties disclosed above and inherent to those skilled in the art pertaining to such inventions. Where the disclosure or subsequently filed claims recite “a” element, “a first” element, or any such equivalent term, the disclosure or claims should be understood to incorporate one or more such elements, neither requiring nor excluding two or more such elements.
- Applicant(s) reserves the right to submit claims directed to combinations and subcombinations of the disclosed inventions that are believed to be novel and non-obvious. Inventions embodied in other combinations and subcombinations of features, functions, elements and/or properties may be claimed through amendment of those claims or presentation of new claims in the present application or in a related application. Such amended or new claims, whether they are directed to the same invention or a different invention and whether they are different, broader, narrower or equal in scope to the original claims, are to be considered within the subject matter of the inventions described herein.
Claims (20)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US14/300,017 US9964764B2 (en) | 2010-04-30 | 2014-06-09 | Visual training devices, systems, and methods |
Applications Claiming Priority (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US32964810P | 2010-04-30 | 2010-04-30 | |
| US12/869,481 US8788197B2 (en) | 2010-04-30 | 2010-08-26 | Visual training devices, systems, and methods |
| US14/300,017 US9964764B2 (en) | 2010-04-30 | 2014-06-09 | Visual training devices, systems, and methods |
Related Parent Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/869,481 Continuation US8788197B2 (en) | 2010-04-30 | 2010-08-26 | Visual training devices, systems, and methods |
| US12/869,487 Continuation US7916310B2 (en) | 2006-10-03 | 2010-08-26 | Measurement apparatus and method |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20140292812A1 true US20140292812A1 (en) | 2014-10-02 |
| US9964764B2 US9964764B2 (en) | 2018-05-08 |
Family
ID=44858946
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/869,481 Active 2031-03-12 US8788197B2 (en) | 2010-04-30 | 2010-08-26 | Visual training devices, systems, and methods |
| US14/300,017 Active 2031-01-30 US9964764B2 (en) | 2010-04-30 | 2014-06-09 | Visual training devices, systems, and methods |
Family Applications Before (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/869,481 Active 2031-03-12 US8788197B2 (en) | 2010-04-30 | 2010-08-26 | Visual training devices, systems, and methods |
Country Status (1)
| Country | Link |
|---|---|
| US (2) | US8788197B2 (en) |
Cited By (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180130371A1 (en) * | 2016-11-09 | 2018-05-10 | Bradley Haber | Digital music reading system and method |
| CN108648555A (en) * | 2018-07-03 | 2018-10-12 | 京东方科技集团股份有限公司 | A kind of intelligent travelling crane training device, system and method |
| US11276242B2 (en) * | 2020-04-06 | 2022-03-15 | David Bakken | Method and system for practicing group formations using augmented reality |
| US11477434B2 (en) | 2018-03-23 | 2022-10-18 | Pcms Holdings, Inc. | Multifocal plane based method to produce stereoscopic viewpoints in a DIBR system (MFP-DIBR) |
| US11689709B2 (en) | 2018-07-05 | 2023-06-27 | Interdigital Vc Holdings, Inc. | Method and system for near-eye focal plane overlays for 3D perception of content on 2D displays |
| US11893755B2 (en) | 2018-01-19 | 2024-02-06 | Interdigital Vc Holdings, Inc. | Multi-focal planes with varying positions |
Families Citing this family (51)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10180572B2 (en) | 2010-02-28 | 2019-01-15 | Microsoft Technology Licensing, Llc | AR glasses with event and user action control of external applications |
| US9229227B2 (en) | 2010-02-28 | 2016-01-05 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a light transmissive wedge shaped illumination system |
| US20150309316A1 (en) | 2011-04-06 | 2015-10-29 | Microsoft Technology Licensing, Llc | Ar glasses with predictive control of external device based on event input |
| US20120249797A1 (en) | 2010-02-28 | 2012-10-04 | Osterhout Group, Inc. | Head-worn adaptive display |
| US9128281B2 (en) | 2010-09-14 | 2015-09-08 | Microsoft Technology Licensing, Llc | Eyepiece with uniformly illuminated reflective display |
| US9129295B2 (en) | 2010-02-28 | 2015-09-08 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear |
| KR20130000401A (en) | 2010-02-28 | 2013-01-02 | 오스터하우트 그룹 인코포레이티드 | Local advertising content on an interactive head-mounted eyepiece |
| US9134534B2 (en) | 2010-02-28 | 2015-09-15 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including a modular image source |
| US9366862B2 (en) | 2010-02-28 | 2016-06-14 | Microsoft Technology Licensing, Llc | System and method for delivering content to a group of see-through near eye display eyepieces |
| US9097891B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment |
| US9223134B2 (en) | 2010-02-28 | 2015-12-29 | Microsoft Technology Licensing, Llc | Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses |
| US9341843B2 (en) | 2010-02-28 | 2016-05-17 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a small scale image source |
| US9091851B2 (en) | 2010-02-28 | 2015-07-28 | Microsoft Technology Licensing, Llc | Light control in head mounted displays |
| US9097890B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | Grating in a light transmissive illumination system for see-through near-eye display glasses |
| US9182596B2 (en) | 2010-02-28 | 2015-11-10 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light |
| US9285589B2 (en) * | 2010-02-28 | 2016-03-15 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered control of AR eyepiece applications |
| US9759917B2 (en) * | 2010-02-28 | 2017-09-12 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered AR eyepiece interface to external devices |
| US8788197B2 (en) * | 2010-04-30 | 2014-07-22 | Ryan Fink | Visual training devices, systems, and methods |
| US8217856B1 (en) * | 2011-07-27 | 2012-07-10 | Google Inc. | Head-mounted display that displays a visual representation of physical interaction with an input interface located outside of the field of view |
| CN115167675A (en) | 2011-09-19 | 2022-10-11 | 视力移动技术有限公司 | Augmented reality device |
| US8941561B1 (en) * | 2012-01-06 | 2015-01-27 | Google Inc. | Image capture |
| US20150202493A1 (en) * | 2012-05-22 | 2015-07-23 | Joseph P. Gannon | System and method for evaluating the physical skills of a person |
| US9429912B2 (en) * | 2012-08-17 | 2016-08-30 | Microsoft Technology Licensing, Llc | Mixed reality holographic object development |
| US20140071163A1 (en) * | 2012-09-11 | 2014-03-13 | Peter Tobias Kinnebrew | Augmented reality information detail |
| US9129429B2 (en) | 2012-10-24 | 2015-09-08 | Exelis, Inc. | Augmented reality on wireless mobile devices |
| US9412201B2 (en) * | 2013-01-22 | 2016-08-09 | Microsoft Technology Licensing, Llc | Mixed reality filtering |
| US20140223647A1 (en) * | 2013-02-12 | 2014-08-14 | Rodney Wayne Long | Goggles for Use in Water |
| KR20140112913A (en) * | 2013-03-14 | 2014-09-24 | 삼성전자���식회사 | Method for displaying dynamic image and an electronic device thereof |
| US9851803B2 (en) * | 2013-03-15 | 2017-12-26 | Eyecam, LLC | Autonomous computing and telecommunications head-up displays glasses |
| GB201310367D0 (en) | 2013-06-11 | 2013-07-24 | Sony Comp Entertainment Europe | Head-mountable apparatus and systems |
| WO2015032824A1 (en) * | 2013-09-04 | 2015-03-12 | Essilor International (Compagnie Generale D'optique) | Methods and systems for augmented reality |
| TWI510811B (en) * | 2013-09-13 | 2015-12-01 | Quanta Comp Inc | Head mounted system |
| CN103591958B (en) * | 2013-11-12 | 2017-01-04 | 中国科学院深圳先进技术研究院 | A kind of worker navigation system based on intelligent glasses and method |
| CN103591951B (en) * | 2013-11-12 | 2017-06-13 | 中国科学院深圳先进技术研究院 | A kind of indoor navigation system and method |
| EP2887322B1 (en) * | 2013-12-18 | 2020-02-12 | Microsoft Technology Licensing, LLC | Mixed reality holographic object development |
| JP2015118578A (en) * | 2013-12-18 | 2015-06-25 | マイクロソフト コーポレーション | Augmented reality information detail |
| CN103761085B (en) * | 2013-12-18 | 2018-01-19 | 微软技术许可有限责任公司 | Mixed reality holographic object is developed |
| US10067341B1 (en) | 2014-02-04 | 2018-09-04 | Intelligent Technologies International, Inc. | Enhanced heads-up display system |
| CN104061925B (en) * | 2014-04-22 | 2017-03-08 | 中国科学院深圳先进技术研究院 | indoor navigation system based on intelligent glasses |
| US10012506B1 (en) | 2014-06-01 | 2018-07-03 | DNP Technology | Navigation guidance system and method of use |
| US9298010B2 (en) * | 2014-08-08 | 2016-03-29 | Marissa J. Sundquist | Wearable optical display with audio functionality |
| EP3007029B1 (en) | 2014-10-07 | 2017-12-27 | LG Electronics Inc. | Mobile terminal and wearable device |
| US9931266B2 (en) | 2015-01-30 | 2018-04-03 | Magno Processing Systems, Inc. | Visual rehabilitation systems and methods |
| WO2016209162A1 (en) * | 2015-06-25 | 2016-12-29 | Razer (Asia-Pacific) Pte. Ltd. | Mobile devices, methods for controlling a mobile device, and computer readable media |
| US10113877B1 (en) * | 2015-09-11 | 2018-10-30 | Philip Raymond Schaefer | System and method for providing directional information |
| CA2948761A1 (en) | 2015-11-23 | 2017-05-23 | Wal-Mart Stores, Inc. | Virtual training system |
| CN105561543B (en) * | 2016-02-04 | 2018-01-02 | 京东方科技集团股份有限公司 | Underwater glasses and its control method |
| US10373391B1 (en) | 2018-04-23 | 2019-08-06 | AbdurRahman Bin Shahzad Bhatti | Augmented reality system for fitness |
| CN109091837B (en) * | 2018-09-06 | 2020-07-10 | 苏州众创时代科创中心 | Football training method |
| US11568640B2 (en) | 2019-09-30 | 2023-01-31 | Lenovo (Singapore) Pte. Ltd. | Techniques for providing vibrations at headset |
| US11144759B1 (en) * | 2020-05-12 | 2021-10-12 | Lenovo (Singapore) Pte. Ltd. | Presentation of graphical objects on display based on input from rear-facing camera |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6989831B2 (en) * | 1999-03-15 | 2006-01-24 | Information Decision Technologies, Llc | Method for simulating multi-layer obscuration from a viewpoint |
| US20080036653A1 (en) * | 2005-07-14 | 2008-02-14 | Huston Charles D | GPS Based Friend Location and Identification System and Method |
| US7751122B2 (en) * | 2005-02-10 | 2010-07-06 | Lumus Ltd. | Substrate-guided optical device particularly for vision enhanced optical systems |
| US8788197B2 (en) * | 2010-04-30 | 2014-07-22 | Ryan Fink | Visual training devices, systems, and methods |
| US8965460B1 (en) * | 2004-01-30 | 2015-02-24 | Ip Holdings, Inc. | Image and augmented reality based networks using mobile devices and intelligent electronic glasses |
Family Cites Families (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7786424B2 (en) * | 2005-05-30 | 2010-08-31 | Andreas Durner | Electronic day and night vision goggles having dual camera |
| WO2008076774A2 (en) * | 2006-12-14 | 2008-06-26 | Oakley, Inc. | Wearable high resolution audio visual interface |
| EP3267228B1 (en) * | 2007-02-16 | 2020-05-27 | NIKE Innovate C.V. | Real-time comparison of athletic information |
| KR20100016169A (en) * | 2007-04-03 | 2010-02-12 | 휴먼 네트워크 랩스, 아이엔씨. | Method and apparatus for acquiring local position and overlaying information |
| US20090005018A1 (en) * | 2007-06-28 | 2009-01-01 | Apple Inc. | Route Sharing and Location |
-
2010
- 2010-08-26 US US12/869,481 patent/US8788197B2/en active Active
-
2014
- 2014-06-09 US US14/300,017 patent/US9964764B2/en active Active
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6989831B2 (en) * | 1999-03-15 | 2006-01-24 | Information Decision Technologies, Llc | Method for simulating multi-layer obscuration from a viewpoint |
| US8965460B1 (en) * | 2004-01-30 | 2015-02-24 | Ip Holdings, Inc. | Image and augmented reality based networks using mobile devices and intelligent electronic glasses |
| US7751122B2 (en) * | 2005-02-10 | 2010-07-06 | Lumus Ltd. | Substrate-guided optical device particularly for vision enhanced optical systems |
| US20080036653A1 (en) * | 2005-07-14 | 2008-02-14 | Huston Charles D | GPS Based Friend Location and Identification System and Method |
| US8788197B2 (en) * | 2010-04-30 | 2014-07-22 | Ryan Fink | Visual training devices, systems, and methods |
Cited By (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180130371A1 (en) * | 2016-11-09 | 2018-05-10 | Bradley Haber | Digital music reading system and method |
| US11893755B2 (en) | 2018-01-19 | 2024-02-06 | Interdigital Vc Holdings, Inc. | Multi-focal planes with varying positions |
| US11477434B2 (en) | 2018-03-23 | 2022-10-18 | Pcms Holdings, Inc. | Multifocal plane based method to produce stereoscopic viewpoints in a DIBR system (MFP-DIBR) |
| US12238270B2 (en) | 2018-03-23 | 2025-02-25 | Interdigital Vc Holdings, Inc. | Multifocal plane based method to produce stereoscopic viewpoints in a DIBR system (MFP-DIBR) |
| CN108648555A (en) * | 2018-07-03 | 2018-10-12 | 京东方科技集团股份有限公司 | A kind of intelligent travelling crane training device, system and method |
| US11689709B2 (en) | 2018-07-05 | 2023-06-27 | Interdigital Vc Holdings, Inc. | Method and system for near-eye focal plane overlays for 3D perception of content on 2D displays |
| US12047552B2 (en) | 2018-07-05 | 2024-07-23 | Interdigital Vc Holdings, Inc. | Method and system for near-eye focal plane overlays for 3D perception of content on 2D displays |
| US12407806B2 (en) | 2018-07-05 | 2025-09-02 | Interdigital Vc Holdings, Inc. | Method and system for near-eye focal plane overlays for 3D perception of content on 2D displays |
| US11276242B2 (en) * | 2020-04-06 | 2022-03-15 | David Bakken | Method and system for practicing group formations using augmented reality |
Also Published As
| Publication number | Publication date |
|---|---|
| US8788197B2 (en) | 2014-07-22 |
| US9964764B2 (en) | 2018-05-08 |
| US20110270522A1 (en) | 2011-11-03 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9964764B2 (en) | Visual training devices, systems, and methods | |
| CN114095622B (en) | System and method for synchronizing image sensors | |
| US10685456B2 (en) | Peer to peer remote localization for devices | |
| JP6391685B2 (en) | Orientation and visualization of virtual objects | |
| US10410562B2 (en) | Image generating device and image generating method | |
| US8914472B1 (en) | Experience sharing for training | |
| US11503265B2 (en) | Rendering extended video in virtual reality | |
| US10908428B2 (en) | Multiple-device system with multiple power and data configurations | |
| JP2021103303A (en) | Information processing device and image display method | |
| US20140333773A1 (en) | Portable audio/ video mask | |
| CN105301771A (en) | Head mounted display, detection device, control method for head mounted display, and computer program | |
| US10536666B1 (en) | Systems and methods for transmitting aggregated video data | |
| JP6384131B2 (en) | Head-mounted display device, control method therefor, and computer program | |
| EP4356692A1 (en) | Systems and methods for lighting subjects for artificial reality scenes | |
| KR20200115631A (en) | Multi-viewing virtual reality user interface | |
| JP2016131782A (en) | Head wearable display device, detection device, control method for head wearable display device, and computer program | |
| WO2023147038A1 (en) | Systems and methods for predictively downloading volumetric data | |
| EP4548150A1 (en) | Compact head-mounted augmented reality system | |
| US20210354035A1 (en) | Interaction in a multi-user environment | |
| US12143561B2 (en) | Image generation system and method | |
| EP2019329A1 (en) | Integrative spectacle-shaped stereo-video multimedia device | |
| US10979733B1 (en) | Systems and methods for measuring image quality based on an image quality metric | |
| Surale et al. | A design of low cost head-mounted display using android smartphone | |
| Ritsos | Architectures for untethered augmented reality using wearable computers | |
| US12536750B1 (en) | Systems and methods for streaming artificial reality data |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: ATHEER LABS, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:FINK, RYAN;REEL/FRAME:039569/0554 Effective date: 20151116 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| AS | Assignment |
Owner name: COTA CAPITAL MASTER FUND, L.P., AS COLLATERAL AGEN Free format text: SECURITY INTEREST;ASSIGNOR:ATHEER, INC.;REEL/FRAME:048154/0759 Effective date: 20181119 Owner name: COTA CAPITAL MASTER FUND, L.P., AS COLLATERAL AGENT, CALIFORNIA Free format text: SECURITY INTEREST;ASSIGNOR:ATHEER, INC.;REEL/FRAME:048154/0759 Effective date: 20181119 |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |
|
| AS | Assignment |
Owner name: ATHEER, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:FINK, RYAN;REEL/FRAME:058637/0323 Effective date: 20220110 Owner name: ATHEER, INC., CALIFORNIA Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE NAME PREVIOUSLY RECORDED ON REEL 039569 FRAME 0554. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:FINK, RYAN;REEL/FRAME:058707/0125 Effective date: 20151116 |
|
| AS | Assignment |
Owner name: WEST TEXAS TECHNOLOGY PARTNERS, LLC, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ATHEER, INC.;REEL/FRAME:058962/0067 Effective date: 20220130 Owner name: WEST TEXAS TECHNOLOGY PARTNERS, LLC, TEXAS Free format text: ASSIGNMENT OF ASSIGNOR'S INTEREST;ASSIGNOR:ATHEER, INC.;REEL/FRAME:058962/0067 Effective date: 20220130 |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 8 |