Dr. Gregory F. Welch – Patents
2020 |
|
![]() | Gregory Welch; Joseph LaViola Jr.; Francisco Guido-Sanz; Gerd Bruder; Mindi Anderson; Ryan Schubert Adaptive Visual Overlay Wound Simulation Patent US 10,854,098 B1, 2020. @patent{Welch2020c, title = {Adaptive Visual Overlay Wound Simulation}, author = {Gregory Welch and Joseph LaViola Jr. and Francisco Guido-Sanz and Gerd Bruder and Mindi Anderson and Ryan Schubert}, url = {https://sreal.ucf.edu/wp-content/uploads/2020/12/US10854098.pdf http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=/netahtml/PTO/srchnum.htm&r=1&f=G&l=50&s1=10,854,098}, year = {2020}, date = {2020-12-01}, number = {US 10,854,098 B1}, abstract = {A wound simulation unit is a physical device designed to help simulate a wound on an object (e.g., a human being or human surrogate such as a medical manikin) for instructing a trainee to learn or practice wound-related treatment skills. For the trainee, the simulation looks like a real wound when viewed using an Augmented Reality (AR) system. Responsive to a change in the anatomic state of the object (e.g., bending a knee or raising o f an arm) not only the spatial location and orientation of the wound stays locked on the object in the AR system, but the characteristics of the wound change based on the physiologic logic o f changing said anatomical state (e.g., greater or less blood flow, opening or closing of the wound).}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A wound simulation unit is a physical device designed to help simulate a wound on an object (e.g., a human being or human surrogate such as a medical manikin) for instructing a trainee to learn or practice wound-related treatment skills. For the trainee, the simulation looks like a real wound when viewed using an Augmented Reality (AR) system. Responsive to a change in the anatomic state of the object (e.g., bending a knee or raising o f an arm) not only the spatial location and orientation of the wound stays locked on the object in the AR system, but the characteristics of the wound change based on the physiologic logic o f changing said anatomical state (e.g., greater or less blood flow, opening or closing of the wound). |
![]() | Gregory Welch; Joseph LaViola Jr.; Francisco Guido-Sanz; Gerd Bruder; Mindi Anderson; Ryan Schubert Multisensory Wound Simulation Patent US 10,803,761 B2, 2020. @patent{Welch2020b, title = {Multisensory Wound Simulation}, author = {Gregory Welch and Joseph LaViola Jr. and Francisco Guido-Sanz and Gerd Bruder and Mindi Anderson and Ryan Schubert}, url = {http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=/netahtml/PTO/srchnum.htm&r=1&f=G&l=50&s1=10803761.PN.&OS=PN/10803761&RS=PN/10803761 https://sreal.ucf.edu/wp-content/uploads/2020/10/welch2020b.pdf}, year = {2020}, date = {2020-10-13}, number = {US 10,803,761 B2}, abstract = {A Tactile-Visual Wound (TVW) simulation unit is a physical device designed to help simulate a wound on a human being or human surrogate (e.g., a medical manikin) for instructing a trainee to learn or practice wound-related treatment skills. For the trainee, the TVW would feel (to the touch) like a real wound, look like a real wound when viewed using an Augmented Reality (AR) system, and appear to behave like a real wound when manipulated.}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A Tactile-Visual Wound (TVW) simulation unit is a physical device designed to help simulate a wound on a human being or human surrogate (e.g., a medical manikin) for instructing a trainee to learn or practice wound-related treatment skills. For the trainee, the TVW would feel (to the touch) like a real wound, look like a real wound when viewed using an Augmented Reality (AR) system, and appear to behave like a real wound when manipulated. |
2019 |
|
![]() | Gregory Welch; Karen Aroian; Steven Talbert; Kelly Allred; Patricia Weinstein; Arjun Nagendran; Remo Pillat Physical-Virtual Patient Bed System Patent US10410541B2, 2019, (Filed: 2017-06-02). BibTeX | Links: @patent{Welch2019ac, title = {Physical-Virtual Patient Bed System}, author = {Gregory Welch and Karen Aroian and Steven Talbert and Kelly Allred and Patricia Weinstein and Arjun Nagendran and Remo Pillat}, url = {https://patents.google.com/patent/US10410541B2/en?oq=10%2c410%2c541 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO2&Sect2=HITOFF&p=1&u=%2Fnetahtml%2FPTO%2Fsearch-bool.html&r=1&f=G&l=50&co1=AND&d=PTXT&s1=10410541&OS=10410541&RS=10410541 https://sreal.ucf.edu/wp-content/uploads/2020/04/Welch2019ac.pdf }, year = {2019}, date = {2019-09-10}, number = {US10410541B2}, location = {US}, note = {Filed: 2017-06-02}, keywords = {}, pubstate = {published}, tppubtype = {patent} } |
![]() | Gregory Welch; Arjun Nagendran; Mary Lou Sole; Laura Gonzalez Physical-Virtual Patient Bed System Patent US10380921B2, 2019, (Filed: 2015-07-01). @patent{Welch2019c, title = {Physical-Virtual Patient Bed System}, author = {Gregory Welch and Arjun Nagendran and Mary Lou Sole and Laura Gonzalez}, url = {https://patents.google.com/patent/US10380921B2/en?q=Physical-Virtual&q=Patient&q=Bed&q=System&oq=Physical-Virtual+Patient+Bed+System http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=10380921.PN.&OS=PN/10380921&RS=PN/10380921 }, year = {2019}, date = {2019-08-13}, number = {US10380921B2}, location = {3100 Technology Parkway}, abstract = {A patient simulation system for healthcare training is provided. The system includes one or more interchangeable shells comprising a physical anatomical model of at least a portion of a patient's body, the shell adapted to be illuminated from behind to provide one or more dynamic images viewable on the outer surface of the shells; a support system adapted to receive the shells via a mounting system, wherein the system comprises one or more image units adapted to render the one or more dynamic images viewable on the outer surface of the shells; one or more interface devices located about the patient shells to receive input and provide output; and one or more computing units in communication with the image units and interface devices, the computing units adapted to provide an interactive simulation for healthcare training.}, note = {Filed: 2015-07-01}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A patient simulation system for healthcare training is provided. The system includes one or more interchangeable shells comprising a physical anatomical model of at least a portion of a patient's body, the shell adapted to be illuminated from behind to provide one or more dynamic images viewable on the outer surface of the shells; a support system adapted to receive the shells via a mounting system, wherein the system comprises one or more image units adapted to render the one or more dynamic images viewable on the outer surface of the shells; one or more interface devices located about the patient shells to receive input and provide output; and one or more computing units in communication with the image units and interface devices, the computing units adapted to provide an interactive simulation for healthcare training. |
![]() | Henry Fuchs; Mingsong Dou; Gregory Welch; Jan-Michael Frahm Methods, systems, and computer readable media for unified scene acquisition and pose tracking in a wearable display Patent US10365711B2, 2019, (Filed: 2013-05-17). @patent{Fuchs2019, title = {Methods, systems, and computer readable media for unified scene acquisition and pose tracking in a wearable display}, author = {Henry Fuchs and Mingsong Dou and Gregory Welch and Jan-Michael Frahm}, url = {https://patents.google.com/patent/US10365711B2/ http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=10,365,711.PN.&OS=PN/10,365,711&RS=PN/10,365,711 }, year = {2019}, date = {2019-07-30}, number = {US10365711B2}, abstract = {Methods, systems, and computer readable media for unified scene acquisition and pose tracking in a wearable display are disclosed. According to one aspect, a system for unified scene acquisition and pose tracking in a wearable display includes a wearable frame configured to be worn by a user. Mounted on the frame are: at least one sensor for acquiring scene information for a real scene proximate to the user, the scene information including images and depth information; a pose tracker for estimating the user's head pose based on the acquired scene information; a rendering unit for generating a virtual reality (VR) image based on the acquired scene information and estimated head pose; and at least one display for displaying to the user a combination of the generated VR image and the scene proximate to the user.}, note = {Filed: 2013-05-17}, keywords = {}, pubstate = {published}, tppubtype = {patent} } Methods, systems, and computer readable media for unified scene acquisition and pose tracking in a wearable display are disclosed. According to one aspect, a system for unified scene acquisition and pose tracking in a wearable display includes a wearable frame configured to be worn by a user. Mounted on the frame are: at least one sensor for acquiring scene information for a real scene proximate to the user, the scene information including images and depth information; a pose tracker for estimating the user's head pose based on the acquired scene information; a rendering unit for generating a virtual reality (VR) image based on the acquired scene information and estimated head pose; and at least one display for displaying to the user a combination of the generated VR image and the scene proximate to the user. |
![]() | Henry Fuchs; Gregory Welch Methods, systems, and computer readable media for improved illumination of spatial augmented reality objects Patent US10321107B2, 2019, (Filed: 2014-11-12). @patent{Fuchs2019b, title = {Methods, systems, and computer readable media for improved illumination of spatial augmented reality objects }, author = {Henry Fuchs and Gregory Welch}, url = {https://patents.google.com/patent/US10321107B2/ http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=10321107.PN.&OS=PN/10321107&RS=PN/10321107 }, year = {2019}, date = {2019-06-11}, number = {US10321107B2}, abstract = {A system for illuminating a spatial augmented reality object includes an augmented reality object including a projection surface having a plurality of apertures formed through the projection surface. The system further includes a lenslets layer including a plurality of lenslets and conforming to curved regions of the of the projection surface for directing light through the apertures. The system further includes a camera for measuring ambient illumination in an environment of the projection surface. The system further includes a projected image illumination adjustment module for adjusting illumination of a captured video image. The system further includes a projector for projecting the illumination adjusted captured video image onto the projection surface via the lenslets layer and the apertures.}, note = {Filed: 2014-11-12}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A system for illuminating a spatial augmented reality object includes an augmented reality object including a projection surface having a plurality of apertures formed through the projection surface. The system further includes a lenslets layer including a plurality of lenslets and conforming to curved regions of the of the projection surface for directing light through the apertures. The system further includes a camera for measuring ambient illumination in an environment of the projection surface. The system further includes a projected image illumination adjustment module for adjusting illumination of a captured video image. The system further includes a projector for projecting the illumination adjusted captured video image onto the projection surface via the lenslets layer and the apertures. |
2017 |
|
![]() | Gregory Welch; Arjun Nagendran; Jason Hochreiter; Laura Gonzalez; Hassan Foroosh System for Detecting Sterile Field Events and Related Methods Patent US 9808549B2, 2017, (Filed: 2015-12-23). @patent{Welch2017ad, title = {System for Detecting Sterile Field Events and Related Methods}, author = {Gregory Welch and Arjun Nagendran and Jason Hochreiter and Laura Gonzalez and Hassan Foroosh}, url = {https://patents.google.com/patent/US9808549B2/en?oq=US+9808549 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=/netahtml/PTO/srchnum.htm&r=1&f=G&l=50&s1=9,808,549.PN.&OS=PN/9,808,549&RS=PN/9,808,549}, year = {2017}, date = {2017-11-07}, number = {US 9808549B2}, location = {US}, abstract = {A system is for monitoring a sterile field associated with a medical procedure. The system may include a sensor being adjacent an area where the medical procedure is to be performed. The sensor is configured to monitor at least the area, a patient for the medical procedure, and a medical technician for the medical procedure. The system may include a processor coupled to the sensor and configured to detect a sterile field event, and an associated location for the sterile field event, and an output device coupled to the processor and configured to generate an alert indicator when the sterile field event is detected, the alert indicator also including the associated location.}, note = {Filed: 2015-12-23}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A system is for monitoring a sterile field associated with a medical procedure. The system may include a sensor being adjacent an area where the medical procedure is to be performed. The sensor is configured to monitor at least the area, a patient for the medical procedure, and a medical technician for the medical procedure. The system may include a processor coupled to the sensor and configured to detect a sterile field event, and an associated location for the sterile field event, and an output device coupled to the processor and configured to generate an alert indicator when the sterile field event is detected, the alert indicator also including the associated location. |
![]() | Gregory Welch; Kurtis Keller; Andrei State; Henry Fuchs; Ryan Schubert Methods, systems, and computer readable media for utilizing synthetic animatronics Patent US 9792715B2, 2017, (Filed: 2013-05-17). @patent{Welch2017ac, title = {Methods, systems, and computer readable media for utilizing synthetic animatronics}, author = {Gregory Welch and Kurtis Keller and Andrei State and Henry Fuchs and Ryan Schubert}, url = {https://patents.google.com/patent/US9792715B2/en?oq=US+9792715 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=9,792,715.PN.&OS=PN/9,792,715&RS=PN/9,792,715}, year = {2017}, date = {2017-10-17}, number = {US 9792715B2}, location = {US}, abstract = {Methods, systems, and computer readable media for utilizing synthetic animatronics are disclosed. According to one aspect, a method for synthetic animatronics includes providing a display surface having different regions that accommodate different positions or deformations of a subject, mapping images of the subject to the different regions on the display surface, and displaying the mapped images on the different regions of the display surface at different times in accordance with a desired animation of the subject.}, note = {Filed: 2013-05-17}, keywords = {}, pubstate = {published}, tppubtype = {patent} } Methods, systems, and computer readable media for utilizing synthetic animatronics are disclosed. According to one aspect, a method for synthetic animatronics includes providing a display surface having different regions that accommodate different positions or deformations of a subject, mapping images of the subject to the different regions on the display surface, and displaying the mapped images on the different regions of the display surface at different times in accordance with a desired animation of the subject. |
![]() | Gregory Welch; Karen Aroian; Steven Talbert; Kelly Allred; Patricia Weinstein; Arjun Nagendran; Remo Pillat Physical-virtual patient bed system Patent US 9679500B2, 2017, (Filed: 2014-03-12). @patent{Welch2017ab, title = {Physical-virtual patient bed system}, author = {Gregory Welch and Karen Aroian and Steven Talbert and Kelly Allred and Patricia Weinstein and Arjun Nagendran and Remo Pillat}, url = {https://patents.google.com/patent/US9679500B2/en?oq=US+9679500 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=9,679,500.PN.&OS=PN/9,679,500&RS=PN/9,679,500}, year = {2017}, date = {2017-06-13}, number = {US 9679500B2}, location = {US}, abstract = {A patient simulation system for healthcare training is provided. The system includes one or more interchangeable shells comprising a physical anatomical model of at least a portion of a patient's body, the shell adapted to be illuminated from behind to provide one or more dynamic images viewable on the outer surface of the shells; a support system adapted to receive the shells via a mounting system, wherein the system comprises one or more image units adapted to render the one or more dynamic images viewable on the outer surface of the shells; one or more interface devices located about the patient shells to receive input and provide output; and one or more computing units in communication with the image units and interface devices, the computing units adapted to provide an interactive simulation for healthcare training.}, note = {Filed: 2014-03-12}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A patient simulation system for healthcare training is provided. The system includes one or more interchangeable shells comprising a physical anatomical model of at least a portion of a patient's body, the shell adapted to be illuminated from behind to provide one or more dynamic images viewable on the outer surface of the shells; a support system adapted to receive the shells via a mounting system, wherein the system comprises one or more image units adapted to render the one or more dynamic images viewable on the outer surface of the shells; one or more interface devices located about the patient shells to receive input and provide output; and one or more computing units in communication with the image units and interface devices, the computing units adapted to provide an interactive simulation for healthcare training. |
![]() | Gregory Welch; Henry Fuchs; Peter Lincoln; Andrew Nashel; Andrei State Methods, systems, and computer readable media for shader-lamps based physical avatars of real and virtual people Patent US 9538167B2, 2017, (Filed: 2010-03-08). @patent{Welch2017aa, title = {Methods, systems, and computer readable media for shader-lamps based physical avatars of real and virtual people}, author = {Gregory Welch and Henry Fuchs and Peter Lincoln and Andrew Nashel and Andrei State}, url = {https://patents.google.com/patent/US9538167B2/en?oq=US+9538167 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=9538167.PN.&OS=PN/9538167&RS=PN/9538167}, year = {2017}, date = {2017-01-03}, number = {US 9538167B2}, location = {US}, abstract = {Methods, systems, and computer readable media for shader lamps-based avatars of real and virtual people are disclosed. According to one method, shader lamps-based avatars of real and virtual objects are displayed on physical target objects. The method includes obtaining visual information of a source object and generating at least a first data set of pixels representing a texture image of the source object. At least one of a size, shape, position, and orientation of a 3D physical target object are determined. A set of coordinate data associated with various locations on the surface of the target object are also determined. The visual information is mapped to the physical target object. Mapping includes defining a relationship between the first and second sets of data, wherein each element of the first set is related to each element of the second set. The mapped visual information is displayed on the physical target object using a display module, such as one or more projectors located at various positions around the target object.}, note = {Filed: 2010-03-08}, keywords = {}, pubstate = {published}, tppubtype = {patent} } Methods, systems, and computer readable media for shader lamps-based avatars of real and virtual people are disclosed. According to one method, shader lamps-based avatars of real and virtual objects are displayed on physical target objects. The method includes obtaining visual information of a source object and generating at least a first data set of pixels representing a texture image of the source object. At least one of a size, shape, position, and orientation of a 3D physical target object are determined. A set of coordinate data associated with various locations on the surface of the target object are also determined. The visual information is mapped to the physical target object. Mapping includes defining a relationship between the first and second sets of data, wherein each element of the first set is related to each element of the second set. The mapped visual information is displayed on the physical target object using a display module, such as one or more projectors located at various positions around the target object. |
2014 |
|
![]() | Richard Gilson; Greg Welch; Nizam Razack Methods for Electronic Directionality of Deep-Brain Stimulation Patent US 8849408B1, 2014, (Filed: 2013-01-04). @patent{Gilson2014, title = {Methods for Electronic Directionality of Deep-Brain Stimulation}, author = {Richard Gilson and Greg Welch and Nizam Razack}, url = {https://patents.google.com/patent/US8849408B1/en?oq=US+%238%2c849%2c408 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=8849408.PN.&OS=PN/8849408&RS=PN/8849408}, year = {2014}, date = {2014-09-30}, number = {US 8849408B1}, location = {US}, abstract = {Methods, systems and devices to provide correction parameters for implanted electrodes by applying a cathode pulse to a bilateral implanted electrode while providing a synchronized anode on the opposite electrode. The electrical field can be “shaped” over space and time to reach more of the targeted area by selecting various combinations of active contacts. The cathode lead directs the electrical field to the target and the placement and number of anode contacts activated determines the electric field path and rate of dissipation based on vertical and horizontal distance and timing. The correction parameter can be applied to anode and cathode contacts on a single implanted lead. Each lead can have plural anode and cathode contacts each independently controllable. Active anodes and cathodes are statically or dynamically selected to generate a shaped electric field to reach the target.}, note = {Filed: 2013-01-04}, keywords = {}, pubstate = {published}, tppubtype = {patent} } Methods, systems and devices to provide correction parameters for implanted electrodes by applying a cathode pulse to a bilateral implanted electrode while providing a synchronized anode on the opposite electrode. The electrical field can be “shaped” over space and time to reach more of the targeted area by selecting various combinations of active contacts. The cathode lead directs the electrical field to the target and the placement and number of anode contacts activated determines the electric field path and rate of dissipation based on vertical and horizontal distance and timing. The correction parameter can be applied to anode and cathode contacts on a single implanted lead. Each lead can have plural anode and cathode contacts each independently controllable. Active anodes and cathodes are statically or dynamically selected to generate a shaped electric field to reach the target. |
2006 |
|
![]() | Greg Welch; Kok-Lim Low; Ramesh Raskar System and method for animating real objects with projected images Patent US 7068274B2, 2006, (Filed: 2001-08-15). @patent{Welch2006c, title = {System and method for animating real objects with projected images }, author = {Greg Welch and Kok-Lim Low and Ramesh Raskar}, url = {https://patents.google.com/patent/US7068274B2/en?oq=US+7%2c068%2c274 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO2&Sect2=HITOFF&p=1&u=%2Fnetahtml%2FPTO%2Fsearch-bool.html&r=1&f=G&l=50&co1=AND&d=PTXT&s1=7068274.PN.&OS=PN/7068274&RS=PN/7068274}, year = {2006}, date = {2006-06-27}, number = {US 7068274B2}, location = {US}, abstract = {A computer implemented method animates a 3D physical object by first acquiring a 3D graphics model of the object. The model is edited with graphics authoring tools to reflect a desired appearance of the object. The edited model is rendered as an image considering a user location and a location of a virtual light. Then, intensity values of the image are corrected according to an orientation of a surface of the object and a radiance at the surface. The 3D physical object can finally be illuminated with the corrected image to give the 3D physical object the desired appearance under the virtual light when viewed from the user location.}, note = {Filed: 2001-08-15}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A computer implemented method animates a 3D physical object by first acquiring a 3D graphics model of the object. The model is edited with graphics authoring tools to reflect a desired appearance of the object. The edited model is rendered as an image considering a user location and a location of a virtual light. Then, intensity values of the image are corrected according to an orientation of a surface of the object and a radiance at the surface. The 3D physical object can finally be illuminated with the corrected image to give the 3D physical object the desired appearance under the virtual light when viewed from the user location. |
2005 |
|
![]() | Ramesh Raskar; Gregory F. Welch; Kok-Lim Low System and method for registering multiple images with three-dimensional objects Patent US 630681B2, 2005, (Filed: 2014-08-14). @patent{Raskar2005, title = {System and method for registering multiple images with three-dimensional objects }, author = {Ramesh Raskar and Gregory F. Welch and Kok-Lim Low}, url = {https://patents.google.com/patent/US6930681B2/en?oq=US+6%2c930%2c681 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO2&Sect2=HITOFF&p=1&u=%2Fnetahtml%2FPTO%2Fsearch-bool.html&r=1&f=G&l=50&co1=AND&d=PTXT&s1=6,930,681.PN.&OS=PN/6,930,681&RS=PN/6930681}, year = {2005}, date = {2005-08-16}, number = {US 630681B2}, location = {US}, abstract = {A computer implemented method registers an image with a 3D physical object by first acquiring a 3D graphics model of an object. Multiple 3D calibration points a surface of the object and corresponding 3D model calibration points in the 3D graphics model are identified. The object is illuminated with a calibration image using a projector at a fixed location. The calibration image is aligned with each of the 3D calibration points on the surface of the 3D physical object to identify corresponding 2D pixels in the calibration image, and then a transformation between the 2D calibration pixels and the corresponding 3D model calibration points is determined to register the projector with the 3D physical object.}, note = {Filed: 2014-08-14}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A computer implemented method registers an image with a 3D physical object by first acquiring a 3D graphics model of an object. Multiple 3D calibration points a surface of the object and corresponding 3D model calibration points in the 3D graphics model are identified. The object is illuminated with a calibration image using a projector at a fixed location. The calibration image is aligned with each of the 3D calibration points on the surface of the 3D physical object to identify corresponding 2D pixels in the calibration image, and then a transformation between the 2D calibration pixels and the corresponding 3D model calibration points is determined to register the projector with the 3D physical object. |
2004 |
|
![]() | Ramesh Raskar; Gregory F. Welch; Kok-Lim Low Method for cross-fading intensities of multiple images of a scene for seamless reconstruction Patent US 6677956B2, 2004, (Filed: 2001-08-15). @patent{Raskar2004, title = {Method for cross-fading intensities of multiple images of a scene for seamless reconstruction }, author = {Ramesh Raskar and Gregory F. Welch and Kok-Lim Low}, url = {https://patents.google.com/patent/US6677956B2/en?oq=US+6%2c677%2c956 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=6,677,956.PN.&OS=PN/6,677,956&RS=PN/6,677,956}, year = {2004}, date = {2004-01-13}, number = {US 6677956B2}, location = {US}, abstract = {A computer implemented method cross-fades intensities of a plurality of overlapping images by identifying pixels in a target image that are only produced by a first source image. The weights of all the corresponding pixels in the first source image are set to one. Pixels in a second source images contributing to the target image are similarly identified and set to one. the weight of each remaining pixel in the first and second images is inversely proportional to a distance to a nearest pixel having a weight of one. Then, the first and second source image can be projected to form the target image.}, note = {Filed: 2001-08-15}, keywords = {}, pubstate = {published}, tppubtype = {patent} } A computer implemented method cross-fades intensities of a plurality of overlapping images by identifying pixels in a target image that are only produced by a first source image. The weights of all the corresponding pixels in the first source image are set to one. Pixels in a second source images contributing to the target image are similarly identified and set to one. the weight of each remaining pixel in the first and second images is inversely proportional to a distance to a nearest pixel having a weight of one. Then, the first and second source image can be projected to form the target image. |
1999 |
|
![]() | Henry Fuchs; Mark Livingston; Gary Bishop; Gregory Francis Welch Dynamic generation of imperceptible structured light for tracking and acquisition of three dimensional scene geometry and surface characteristics in interactive three dimensional computer graphics applications Patent US 5870136A, 1999, (Filed: 1997-12-05). @patent{Fuchs1999b, title = {Dynamic generation of imperceptible structured light for tracking and acquisition of three dimensional scene geometry and surface characteristics in interactive three dimensional computer graphics applications}, author = {Henry Fuchs and Mark Livingston and Gary Bishop and Gregory Francis Welch}, url = {https://patents.google.com/patent/US5870136A/en?oq=US+5%2c870%2c136 http://patft.uspto.gov/netacgi/nph-Parser?Sect1=PTO1&Sect2=HITOFF&d=PALL&p=1&u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&r=1&f=G&l=50&s1=5,870,136.PN.&OS=PN/5,870,136&RS=PN/5,870,136}, year = {1999}, date = {1999-02-09}, number = {US 5870136A}, location = {US}, abstract = {Methods, systems and computer products are provided for tracking objects within a scene using imperceptible structured light. The imperceptible structured light may be used in environments where humans work and therefore avoids disorienting observers. The structured light patterns are generated dynamically, allowing tracking without physical landmarks. Moreover, the occlusion of the generated landmarks is overcome by determining the position of objects using a determined camera position. The imperceptible structured light may be used in systems without tracking. The tracking may be used without imperceptible structured light.}, note = {Filed: 1997-12-05}, keywords = {}, pubstate = {published}, tppubtype = {patent} } Methods, systems and computer products are provided for tracking objects within a scene using imperceptible structured light. The imperceptible structured light may be used in environments where humans work and therefore avoids disorienting observers. The structured light patterns are generated dynamically, allowing tracking without physical landmarks. Moreover, the occlusion of the generated landmarks is overcome by determining the position of objects using a determined camera position. The imperceptible structured light may be used in systems without tracking. The tracking may be used without imperceptible structured light. |