Search

US-12625543-B2 - Computer system and method employing artificial intelligence engine

US12625543B2US 12625543 B2US12625543 B2US 12625543B2US-12625543-B2

Abstract

A user enters natural language text into a mobile phone. The mobile phone detects features of the user. An artificial reality engine modifies features of the user based on an intent derived from the natural language.

Inventors

  • Vance W. Boatwright

Assignees

  • Vance W. Boatwright

Dates

Publication Date
20260512
Application Date
20240927

Claims (20)

  1. 1 . A computer system comprising: a sensor positioned to capture at least one position of object to generate a digital signal representing the at least one position of the object; a pose estimator connected to the sensor and executable by at least one processor to estimate a pose of the object based on the digital signal; an input device positioned to a capture a user input from a user in the form of language; an artificial intelligence interface module connected to the input device and operable by the at least one processor to transmit the language to an artificial intelligence engine executable by the at least one processor to determine an intent of the user based on the language, and to generate code based on the intent, the artificial intelligence engine entering the intent in a rendering engine connected to the pose estimator and to the artificial intelligence engine and executable by the at least one processor to render the pose in a scene that includes at least the pose and with the scene being modified based on the generated code; a display; a display module connected to the rendering engine to create a visual representation of the scene as modified based on the generated code for displaying by the display; a humanoid robot; and a robot controller connected to the robot and executable by the at least one processor, wherein the artificial intelligence engine controls the robot through the robot controller based on the generated code.
  2. 2 . The computer system of claim 1 , wherein the object is at least one body part of the user.
  3. 3 . The computer system of claim 1 , further comprising: a mobile phone that includes: mobile phone body; and a camera mounted to the mobile phone body, wherein the sensor is a sensor of the camera.
  4. 4 . The computer system of claim 3 , wherein the camera is an optical camera and the sensor is an optical sensor detecting light in the visible spectrum.
  5. 5 . The computer system of claim 1 , wherein the input device is a keyboard.
  6. 6 . The computer system of claim 1 , wherein the input device is a microphone, further comprising: a speech-to-text interface module connected to the microphone and operable by the at least one processor to transmit speech captured by the microphone to a speech-to-text service executable by the at least one processor convert speech captured by the microphone to text, the text being the language used by the artificial intelligence engine.
  7. 7 . The computer system of claim 1 , wherein the language is natural language.
  8. 8 . The computer system of claim 1 , further comprising: a body part tracker connected to the pose estimator and executable by at least one processor to determine body part features of the user.
  9. 9 . The computer system of claim 8 , wherein the body part tracker determined gestures from the body part features of the user, the gestures being used by the artificial intelligence engine to update the scene.
  10. 10 . The computer system of claim 8 , wherein the body part tracker is a hand tracker.
  11. 11 . The computer system of claim 1 , further comprising: an artificial reality generator connected to the artificial intelligence engine and executable by the at least one processor; and a digital universe connected to the artificial reality generator and executable by the at least one processor, the artificial reality generator controlling the digital universe based on the generated code.
  12. 12 . The computer system of claim 1 , further comprising: a virtual objects generator connected to the artificial intelligence engine and executable by the at least one processor to generate virtual objects based on the generated code for rendering by the rendering engine.
  13. 13 . The computer system of claim 12 , further comprising: an object interaction module connected to the virtual objects generator and executable by the at least one processor to detect collisions of the virtual objects.
  14. 14 . A computer-implemented method comprising: capturing, with a sensor, at least one position of object to generate a digital signal representing the at least one position of the object; estimating, with a pose estimator connected to the sensor and executable by at least one processor, a pose of the object based on the digital signal; capturing, with an input device, a user input from a user in the form of language; determining, with an artificial intelligence engine connected to the input device and operable by the at least one processor, an intent of the user based on the language, the artificial intelligence engine generating a code based on the intent; rendering, with a rendering engine connected to the pose estimator and to the artificial intelligence engine and executable by the at least one processor, the pose in a scene that includes at least the pose and with the scene being modified based on the generated code; creating, with a display connected to the rendering engine, a visual representation of the scene as modified based on the generated code; and determining, with a body part tracker connected to the sensor and executable by at least one processor, body part features of the user; wherein the body part tracker determines gestures from the body part features of the user, the gestures being used by the artificial intelligence engine to update the scene.
  15. 15 . The method of claim 14 , wherein the object is at least one body part of the user.
  16. 16 . The method of claim 14 , wherein the sensor is a sensor of the camera.
  17. 17 . The method of claim 16 , wherein the camera is an optical camera and the sensor is an optical sensor detecting light in the visible spectrum.
  18. 18 . The method of claim 14 , wherein the input device is a keyboard.
  19. 19 . The method of claim 14 , wherein the input device is a microphone, further comprising: converting, with a speech-to-text service connected to the microphone and executable by at least one processor, speech captured by the microphone to text, the text being the language used by the artificial intelligence engine.
  20. 20 . The method of claim 14 , wherein the language is natural language.

Description

BACKGROUND OF THE INVENTION 1). Field of the Invention This invention relates to a computer system and a computer implemented method using an artificial intelligence engine. 2). Discussion of Related Art Since their development in the 1940's digital computers have been programmed to carry out very complex tasks, such as discovering proofs for mathematical theorems or playing chess, with great proficiency. Despite continuing advances in computer processing speeds and memory capacity, there are as yet no programs that can match full human flexibility over wider domains or in tasks requiring much everyday knowledge. Some programs have attained the performance levels of human experts and professionals in executing certain specific task so that artificial intelligence, in this limited sense, is found in applications as diverse as medical diagnoses, computer search engines, voice, and handwriting recognition, and chatbots. Further advances in artificial intelligence have given digital computers or computer-controlled robots the ability to perform tasks commonly associated with intelligent beings. The term “artificial intelligence” is frequently used for describing systems that appear to display the intellectual process characteristics of humans, such as the ability to reason, discover meaning, generalize, or learn from past experience. SUMMARY OF THE INVENTION The invention provides a computer system including a sensor positioned to capture at least one position of object to generate a digital signal representing the at least one position of the object, a pose estimator connected to the sensor and executable by at least one processor to estimate a pose of the object based on the digital signal, an input device positioned to a capture a user input from a user in the form of language, an artificial intelligence interface module connected to the input device and operable by the at least one processor to transmit the language to an artificial intelligence engine executable by the at least one processor to determine an intent of the user based on the language, and to generate code based on the intent, the artificial intelligence engine entering the intent in a rendering engine connected to the pose estimator and to the artificial intelligence engine and executable by the at least one processor to render the pose in a scene that includes at least the pose and with the scene being modified based on the generated code, a display, and a display module connected to the rendering engine to create a visual representation of the scene as modified based on the generated code for displaying by the display. The computer system may further include that the object is at least one body part of the user. The computer system may further include a mobile phone that may include a mobile phone body, and a camera mounted to the mobile phone body, wherein the sensor is a sensor of the camera. The computer system may further include that the camera is an optical camera and the sensor is an optical sensor detecting light in the visible spectrum. The computer system may further include that the input device is a keyboard. The computer system may further include that the input device is a microphone, further including a speech-to-text interface module connected to the microphone and operable by the at least one processor to transmit speech captured by the microphone to a speech-to-text service executable by the at least one processor convert speech captured by the microphone to text, the text being the language used by the artificial intelligence engine. The computer system may further include that the language is natural language. The computer system may further include a body part tracker connected to the pose estimator and executable by at least one processor to determine body part features of the user. The computer system may further include that the body part tracker determined gestures from the body part features of the user, the gestures being used by the artificial intelligence engine to update the scene. The computer system may further include that the body part tracker is a hand tracker. The computer system may further include a humanoid robot, and a robot controller connected to the robot and executable by the at least one processor, wherein the artificial intelligence engine controls the robot through the robot controller based on the generated code. The computer system may further include an artificial reality generator connected to the artificial intelligence engine and executable by at least one processor, and a digital universe connected to the artificial reality generator and executable by the at least one processor, the artificial reality generator controlling the digital universe based on the generated code. The computer system may further include a virtual objects generator connected to the artificial intelligence engine and executable by the at least one processor to generate virtual objects based on the generated code fo