Search

US-20260127793-A1 - METHOD FOR DISPLAYING VIRTUAL IMAGE OF DIGITAL ASSISTANT AND RELATED APPARATUS

US20260127793A1US 20260127793 A1US20260127793 A1US 20260127793A1US-20260127793-A1

Abstract

A method for displaying a virtual image of a digital assistant is provided. In this method, the virtual image of the digital assistant includes a plurality of feature objects. After an electronic device obtains a user input, the electronic device generates a corresponding signal instruction, to trigger a change of a form of the virtual image displayed on an interface, so that different working states of the digital assistant are represented by using different forms. In this way, a user determines a current interaction status with the digital assistant based on a change status of a form of the digital assistant, to make interaction between the user and the digital assistant more naturally.

Inventors

  • Yifan Wang
  • Shunji Jiang
  • Li Lin
  • Jiangzhen Zheng
  • Zongbo Wang

Assignees

  • HUAWEI TECHNOLOGIES CO., LTD.

Dates

Publication Date
20260507
Application Date
20251229
Priority Date
20230629

Claims (20)

  1. 1 . A method for displaying a virtual image of a digital assistant implemented by an electronic device, comprising: displaying the virtual image in a first form on a first interface, wherein the virtual image comprises a plurality of feature objects; and detecting a signal instruction generated by the electronic device; and displaying the virtual image in a second form, wherein the signal instruction is generated based on a first user input, at least one of the plurality of feature objects is presented in the first form and the second form, and changing from the first form to the second form means that at least one feature object undergoes at least one of a shape change, a position change, a motion law change, or a display area change.
  2. 2 . The method according to claim 1 , wherein the plurality of feature objects comprise a first feature object, and a shape change of the first feature object comprises at least one of the following changes: scale-up, scale-down, stretching, compression, rotation, and switching between a dominant form, a geometric array, or a long bar.
  3. 3 . The method according to claim 2 , wherein the plurality of feature objects further comprise a second feature object, a dominant form of the first feature object is a ring, the second feature object is a sphere, and a circle center of the ring and a sphere center of the sphere are at a same point, wherein a position change of the first feature object comprises at least one of a trajectory position change after a roll of the first object around the sphere or a position change caused by overall movement or rotation of the virtual image, wherein a motion law of the first feature object comprises that the ring rotates around the sphere according to a preset trajectory by using the circle center as a pivot, and wherein a motion law of the second feature object comprises that the sphere performs reciprocating motion according to a preset law along a straight line that passes through the circle center of the ring and that is perpendicular to a plane on which the ring is located.
  4. 4 . The method according to claim 3 , further comprising: detecting a second user input; and displaying a third feature object on the first feature object, wherein the third feature object is capable of moving on the first feature object and indicating an orientation of a user.
  5. 5 . The method according to claim 1 , wherein the virtual image has a to-be-woken-up state and a listening state, and the first interface is a lock screen interface or any screen interface after unlocking, and wherein the displaying the virtual image in the first form on the first interface comprises: displaying the virtual image on the first interface by using the to-be-woken-up state as the first form, wherein the to-be-woken-up state comprises the first feature object, the first feature object as a navigation prompt line is displayed as the long bar, the navigation prompt line is used as a reference line for screen gesture navigation, and the to-be-woken-up state is persistently displayed on the first interface when no input of waking up the digital assistant is received; or detecting a third user input, and displaying the virtual image on the first interface by using the listening state as the first form, wherein the third user input is used to wake up the digital assistant, the listening state comprises the first feature object and the second feature object, and the first feature object is displayed in the dominant form.
  6. 6 . The method according to claim 5 , wherein the first interface is the lock screen interface or the screen interface after unlocking, and the first form is the listening state; and displaying the virtual image in the first form on the first interface comprises: first displaying the first feature object by using the geometric array; and changing the first feature object from the geometric array to the dominant form, and displaying the second feature object.
  7. 7 . The method according to claim 6 , wherein the first displaying the first feature object by using the geometric array comprises: displaying a geometric figure of a part of the array at an initial display position, and then displaying the geometric figure on an entire screen, wherein the initial display position is determined based on a user input; and when no user input is detected, the initial display position is a region for displaying the virtual image.
  8. 8 . The method according to claim 1 , wherein the displaying the virtual image in the first form on the first interface comprises: displaying the virtual image at a preset position on the first interface; or overlay displaying the virtual image on a target content element in the first interface, wherein the target content element is a content element of a preset type.
  9. 9 . The method according to claim 1 , wherein the first interface is a camera interface, and the target content element is a shutter control in the camera interface.
  10. 10 . The method according to claim 5 , wherein the first interface is the lock screen interface or the screen interface after unlocking, the first form is the to-be-woken-up state, the second form is the listening state, and the signal instruction is generated after a user makes a preset gesture to the digital assistant in the first form,
  11. 11 . The method according to claim 10 , wherein the plurality of feature objects further comprises the second feature object, and wherein in the second form, the first feature object is in the dominant form of the ring, the second feature object is the sphere, and the ring is located on an equator at a top view angle of the sphere.
  12. 12 . The method according to claim 5 , wherein the virtual image further has a task execution state, and the task execution state comprises the first feature object, and wherein on at least one process interface in a process in which the digital assistant executes an automated task, the first feature object in the task execution state correspondingly generates at least one of a shape change and a position change.
  13. 13 . The method according to claim 12 , wherein the first feature object is the geometric array, and a display position of the geometric array matches a target control in the process interface to simulate a tap event of the target control to go to a next process interface.
  14. 14 . The method according to claim 12 , wherein the first form is the listening state, the second form is the task execution state, both the listening state and the task execution state comprise the first feature object, and the first feature object in the listening state and the first feature object in the task execution state have different forms, and wherein in the task execution state, the form of the first feature object comprises: the geometric array on an interface, an outer frame of a dialog card, an outer frame of an input box when input is performed by invoking the digital assistant, or an outer frame of a task result page.
  15. 15 . The method according to claim 14 , wherein the first feature object has different form changes in different task progresses of the automated task.
  16. 16 . The method according to claim 14 , wherein when the automated task is a smart recognition task, and the first feature object changes from a geometric array that is laid out to the entire screen to a geometric array that is displayed only on a recognized executable object.
  17. 17 . The method according to claim 14 , wherein in the task execution state, the form change of the first feature object comprises: the first feature object is first displayed in the dominant form of the ring on a periphery of the second feature object; the first feature object as the ring is then detached from the second feature object, and is scaled up and displayed as the outer frame of the dialog card; the first feature object changes from the outer frame of the dialog card to a geometric array in a dialog bubble; and the virtual image is displayed in the listening state after execution of the automated task is completed.
  18. 18 . The method according to claims 1 wherein the virtual image has a to-be-woken-up state, the to-be-woken-up state at least comprises the first feature object, and the first feature object is displayed as the long bar, and wherein the method further comprises: displaying the second feature object in response to detecting that a focus of a line of sight of a human eye falls in a region in which the screen is located.
  19. 19 . An apparatus for displaying a virtual image of a digital assistant, comprising: a memory storing computer programming instructions; and at least one processor coupled to the memory and configured to execute the computer programming instructions to cause the apparatus to perform operations of: displaying the virtual image in a first form on a first interface, wherein the virtual image comprises a plurality of feature objects; and detecting a signal instruction generated by the electronic device; and displaying the virtual image in a second form, wherein the signal instruction is generated based on a first user input, at least one of the plurality of feature objects is presented in the first form and the second form, and changing from the first form to the second form means that at least one feature object undergoes at least one of a shape change, a position change, a motion law change, or a display area change.
  20. 20 . A non-transitory computer storage medium storing computer programming instructions, that when executed by one or more processors, cause the one or more processors to perform steps of: displaying the virtual image in a first form on a first interface, wherein the virtual image comprises a plurality of feature objects; and detecting a signal instruction generated by the electronic device; and displaying the virtual image in a second form, wherein the signal instruction is generated based on a first user input, at least one of the plurality of feature objects is presented in the first form and the second form, and changing from the first form to the second form means that at least one feature object undergoes at least one of a shape change, a position change, a motion law change, or a display area change.

Description

CROSS REFERENCE TO RELATED APPLICATION This application is a continuation of International Application No. PCT/CN2024/097229, filed on Jun. 4, 2024, which claims priority to Chinese Patent Application No. 202310786258.8, filed on Jun. 29, 2023 and Chinese Patent Application No. 202410178141.6, filed on Feb. 8, 2024. All of the aforementioned patent applications are hereby incorporated by reference in their entireties. TECHNICAL FIELD This application relates to the field of human-machine interaction technologies, and in particular, to a method for displaying a virtual image of a digital assistant and a related apparatus. BACKGROUND With rapid development of smart devices, configuring a digital assistant on a smart device gradually becomes a mainstream. A series of operations of the smart device are automatically completed by using the digital assistant instead of manual operations of a user, so that both hands of the user can be released, thereby improving use experience of the smart device. Currently, when using the digital assistant, the user usually needs to wake up the digital assistant by using a voice instruction or a manual operation, to access a use interface of the digital assistant. An assistant image is displayed on the use interface of the digital assistant. The assistant image is usually a static two-dimensional pattern or a two-dimensional pattern that continuously moves according to a specific law. On the use interface of the digital assistant, after the user further inputs a voice instruction or a gesture instruction, the user usually needs to wait for the digital assistant to output a recognition result, to determine whether the digital assistant has received the voice instruction or the gesture instruction. Then, in a process in which the digital assistant recognizes or processes the instruction, a form of the digital assistant does not change. Therefore, after the user inputs the voice instruction or the gesture instruction, the user usually needs to wait for the digital assistant to output the recognition result, to determine whether the digital assistant has received the voice instruction or the gesture instruction. In other words, when the digital assistant recognizes or processes the instruction input by the user, the user cannot obtain a feedback from the digital assistant. It is difficult for the user to recognize a current working state of the digital assistant. As a result, the digital assistant may mistakenly consider that the digital assistant has not received the instruction, and the user repeatedly delivers the instruction. SUMMARY This application provides a method for displaying a virtual image of a digital assistant, to make interaction between a user and a digital assistant more naturally. A first aspect of this application provides a method for displaying a virtual image of a digital assistant. The method is applied to an electronic device. The virtual image of the digital assistant includes a plurality of feature objects. In the method, the virtual image is first displayed in the first form on the first interface; and then, a signal instruction generated by the electronic device is detected, and the virtual image is displayed in a second form. The signal instruction is generated by the electronic device based on a received first user input. The user input may be before or after the first form is displayed. In other words, after receiving the first user input, the electronic device generates the signal instruction, to trigger the virtual image to switch from the first form to the second form, so that a form of the virtual image displayed on an interface of the electronic device changes. In addition, at least one of the plurality of feature objects is presented in both the first form and the second form. Changing from the first form to the second form means that at least one feature object undergoes at least one of a shape change, a position change, a motion law change, and a display area change. Different forms of the virtual image represent different working states of the digital assistant. In other words, the digital assistant uses the form of the virtual image to indicate a current working state of the digital assistant. In this way, when the form of the virtual image displayed on the interface changes, a user can intuitively and clearly understand the change of the working state of the digital assistant, to clearly know a current response status of the digital assistant to the user instruction. In this solution, it is set that the virtual image of the digital assistant includes the plurality of feature objects. After the electronic device obtains a user input, the electronic device generates a corresponding signal instruction, to trigger a change of the form of the virtual image displayed on the interface, so that different working states of the digital assistant are represented by using different forms. In this way, the user determines a current interaction status wi