US8452602B1 - Structuring verbal commands to allow concatenation in a voice interface in a mobile device - Google Patents
Structuring verbal commands to allow concatenation in a voice interface in a mobile device Download PDFInfo
- Publication number
- US8452602B1 US8452602B1 US13/621,018 US201213621018A US8452602B1 US 8452602 B1 US8452602 B1 US 8452602B1 US 201213621018 A US201213621018 A US 201213621018A US 8452602 B1 US8452602 B1 US 8452602B1
- Authority
- US
- United States
- Prior art keywords
- level
- spoken utterance
- slot
- command format
- computing device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/226—Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics
- G10L2015/228—Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics of application context
Definitions
- a user interface allows a user to interact with a computing device.
- the computing device includes a display that provides a text-based user interface and/or graphical user interface.
- a user can interact with the interface using an input device, such as a mouse or touchpad controlled cursor, a keypad, a track ball, and the like.
- the user can interact with the user interface with audible commands.
- a method is provided.
- a spoken utterance is received at a computing device.
- the spoken utterance includes at least a first level of a multi-level command format.
- the first level relates to an application.
- Speech recognition is performed on the spoken utterance to obtain text corresponding to the spoken utterance.
- the application is identified.
- a representation of the identified application is displayed on a display of the computing device.
- the method further involves determining whether the text includes a second level of the multi-level command format.
- the second level relates to an action. If the text includes the second level of the multi-level command format, the action is identified and the identified action is initiated.
- the computing device waits for a first predetermined period of time to receive an additional spoken utterance that includes the second level of the multi-level command format. If no spoken utterance that includes the second level is received within the first predetermined period of time, at least one of an audible or visual action prompt is provided.
- a computer readable medium has stored thereon instructions that, when executed by a computing device having at least one processor, cause the computing device to perform functions.
- the functions include: receiving a spoken utterance that includes at least a first level of a multi-level command format, in which the first level identifies an application; in response to receiving the spoken utterance, displaying a representation of the identified application on a display of the computing device; if the spoken utterance includes a second level of the multi-level command format, in which the second level identifies an action, initiating the identified action; and if the spoken utterance does not include the second level of the multi-level command format, waiting for a first predetermined period of time to receive an additional spoken utterance that includes the second level of the multi-level command format and, if no spoke utterance that includes the second level is received within the first predetermined period of time, providing at least one of an audible or visual action prompt.
- a computing device includes at least one processor, data storage, and a plurality of instructions stored in the data storage.
- the plurality of instructions are executable by the at least one processor to cause the computing device to perform functions.
- the functions include: receiving a spoken utterance that includes at least a first level of a multi-level command format, in which the first level identifies an application; in response to receiving the spoken utterance, displaying a representation of the identified application on a display of the computing device; if the spoken utterance includes a second level of the multi-level command format, in which the second level identifies an action, initiating the identified action; and if the spoken utterance does not include the second level of the multi-level command format, waiting for a first predetermined period of time to receive an additional spoken utterance that includes the second level of the multi-level command format and, if no utterance that includes the second level is received within the first predetermined period of time, providing at least one of an audible or visual action prompt.
- FIG. 1 is a block diagram of a computing device in accordance with an example embodiment
- FIG. 2 is a flowchart of processes for providing a voice user interface in accordance with an example embodiment
- FIGS. 3A-3F are screenshots of a voice user interface in accordance with an example embodiment.
- FIG. 4 a block diagram showing a structured audible command being provided in the context of a music player application.
- FIG. 5 is a flowchart of a method for analyzing spoken input with reference to a multi-level command format in accordance with an example embodiment.
- the present disclosure is directed generally to user interfaces that process audible inputs to control user interactions with computing devices and, more particularly, to such user interfaces that provide visual cues for audible inputs to control user interactions with computing devices.
- a powerful voice input or command language can be difficult to learn and to provide visual cues therefor.
- a very simple command language for which a device can provide visual and audible cues can be very slow to use for an expert user.
- the present disclosure provides a user interface that is adaptable for both beginner and expert users.
- FIG. 1 is a block diagram of a computing device 20 in accordance with an example embodiment.
- the computing device 20 can be any known computing device that can be configured to processes audible inputs to control user interactions therewith.
- the computing device 20 can be a mobile device, such as a smart phone, a personal digital assistance, a global positioning service device, a tablet computer, a laptop, etc.
- the computing device 20 illustrated in FIG. 1 include a user interface module 22 , a network-communication interface module 24 , one or more processors 26 , and data storage 28 , all of which may be linked together via a system bus, network, or other connection mechanism 30 .
- the user interface module 22 can be operable to send data to and/or receive data from external user input/output devices.
- the user interface module 22 can be configured to send and/or receive data to and/or from user input devices such as a keyboard, a keypad, a touch screen, a computer mouse, a track ball, a joystick, a camera, a voice recognition module, and/or other similar devices.
- the user interface module 22 can also be configured to provide output to user display devices, such as one or more cathode ray tubes (CRT), liquid crystal displays (LCD), light emitting diodes (LEDs), displays using digital light processing (DLP) technology, printers, light bulbs, and/or other similar devices, either now known or later developed.
- CTR cathode ray tubes
- LCD liquid crystal displays
- LEDs light emitting diodes
- DLP digital light processing
- the user interface module 22 can also include a microphone to receive audible input, such as spoken utterances, and can be configured to generate audible outputs to a speaker, speaker jack, audio output port, audio output device, earphones, and/or other similar devices.
- the computing device 20 with the user interface module 22 can be used to display a user interface.
- the network-communications interface module 24 can include one or more wireless interfaces 32 and/or one or more wired interfaces 34 that are configurable to communicate via a communication network to other devices coupled to the network.
- computing device 20 may communicate with a server that is configured to perform speech recognition on spoken utterances received by computing device 20 .
- the wireless interfaces 32 can include one or more wireless transmitters, receivers, and/or transceivers, such as a Bluetooth transceiver, a Zigbee transceiver, a Wi-Fi transceiver, a WiMAX transceiver, and/or other similar type of wireless transceiver configurable to communicate via a wireless network.
- the wired interfaces 34 can include one or more wired transmitters, receivers, and/or transceivers, such as an Ethernet transceiver, a Universal Serial Bus (USB) transceiver, or similar transceiver configurable to communicate via a twisted pair wire, a coaxial cable, a fiber-optic link, or a similar physical connection to a wired network.
- wired transmitters such as an Ethernet transceiver, a Universal Serial Bus (USB) transceiver, or similar transceiver configurable to communicate via a twisted pair wire, a coaxial cable, a fiber-optic link, or a similar physical connection to a wired network.
- USB Universal Serial Bus
- the processors 26 can include one or more general purpose processors and/or one or more special purpose processors, for example, digital signal processors, application specific integrated circuits, and the like.
- the processors 26 can be configured to execute computer-readable program instructions 36 that are contained in the data storage 28 and/or other instructions as described herein.
- the data storage 28 can include one or more computer-readable storage media that can be read and/or accessed by at least one of the processors 26 .
- the one or more computer-readable storage media can include volatile and/or non-volatile storage components, such as optical, magnetic, organic or other memory or disc storage, which can be integrated in whole or in part with at least one of the processors 26 .
- the data storage 28 can be implemented using a single physical device, such as one optical, magnetic, organic or other memory or disc storage unit, while in other embodiments, the data storage can be implemented using two or more physical devices.
- the data storage 28 can include the computer-readable program instructions 36 , actual environment data 38 , and perhaps additional data.
- the actual environment data 38 can include at least some of the data used by one or more processes and/or threads of a software application.
- the data storage 28 can additionally include storage required to perform at least part of the herein-described methods and techniques.
- a flowchart 60 is illustrated that includes processes for providing a voice user interface in accordance with an example embodiment.
- the processes of FIG. 2 may be performed by utilizing various hardware and/or software components of the computing device 20 of FIG. 1 .
- a user can speak an audible command to perform a task.
- Such audible commands can be structured into multi-level trees.
- the command may include an Application level, an Action level, and a Slot level.
- Such a command may include an email Application level, a new message Action level, and one or more Slot levels for information associated with a recipient, a subject line, a message, etc.
- the command may include an Application level and a Slot level.
- Such a command may include a calculator Application level and one or more Slot levels for numbers and mathematical operators.
- the command may include as many or few levels as necessary or desired.
- the flowchart 60 begins at a block 62 to identify an application or a plurality of applications that can be executed on a computing device.
- the block 62 identifies an application that can be accessed by a user with an Application level audible input, such as a voice command received by a microphone of the computing device.
- the block 62 identifies a plurality of applications that can each be accessed by a user with one or more of a plurality of different Application level audible inputs.
- Such text is displayed by a block 66 , for example, on a display of the computing device.
- a displayed text word or string represents each of the one or more applications identified by the block 62 .
- the block 66 can display a plurality of text words or strings that each can represent one of a plurality of applications.
- the displayed text words or strings are displayed on the computing device to provide a visual cue of Application level audible inputs or commands to access such applications.
- the display of the computing device can include a plurality of applications, of which, some may not be accessible by an audible input. Further, such list of applications may be displayed in any suitable format, for example, in a scrollable list, a 3-D display, a spiral list, etc.
- such inputs include an Application level audible input or a manual selection to access an application.
- a user can access the application by speaking the displayed text of the Application level audible input or by manually selecting the application, such as by pressing on a touchpad that overlays the displayed text. If the block 68 receives any of the Application level audible input or manual selection, then the block 68 accesses the application. Thereafter, the display of the computing device can be automatically updated with a user interface for the application.
- a block 70 identifies an action and/or a slot associated with the application. Further, in the present example, the block 70 identifies one or more actions and/or slots that can be performed or accessed by an Action/Slot level audible input.
- an action can be a sub-application or command within the application and a slot can be an information box or cell.
- an application can be an email application, an action can be to send a new email, and a slot can include the recipient of the email.
- a single application may include one or more actions and one or more slots for each action.
- An application may also only include one or more slots, for example, an application can be a calculator application and the slots can include numbers and mathematical operators.
- a block 72 determines text of the Action/Slot level audible input for performing or accessing such application and/or slot.
- a block 74 displays such text on the display of the computing device.
- text of Action/Slot level audible commands for a plurality actions and/or slots is displayed automatically upon the block 68 receiving the audible input to access an application.
- the text of Action/Slot level audible commands for a plurality of actions and/or slots can be displayed after an active user command to process the input of the block 68 .
- Such active user command may include pressing a button to process speech, for example.
- such inputs include an Action/Slot level audible input or a manual selection to perform an action and/or to access and input information into a slot.
- a user can perform the action or access the slot by speaking the displayed text of the Action/Slot level audible input or by manually selecting the action or slot, such as by pressing on a touchpad that overlays the displayed text.
- the block 76 receives any of the Action/Slot level audible input or manual selection, then the block 76 performs the requested interaction, for example, initiate a new email message or enter a mathematical formula. Thereafter, the display of the computing device can be automatically updated in accordance with the requested interaction.
- the processes of the block 76 can be repeated for any number of actions/slots, for example, to allow a user to continue issuing audible commands to prepare and send an email to a recipient.
- the inputs received by the block 76 can cause the display of the computing device to be automatically updated in response to each audible and/or manual input.
- the method 60 also includes a decision block 78 , which identifies whether there is a pause for a predetermined time before an input is received.
- the block 78 determines if there is about a 0.5-1.5 second pause after an Action level audible input. If there is no pause, then control passes back to the processing of the block 76 , as described above. If there is a pause, then control passes to a block 80 to prompt the user to provide an input.
- the prompt can be a visual cue, such as highlighting an action or a slot, or an audible cue, such as a beep or an audible command to select an action or input information into a slot. After the user is prompted by the block 80 , control can pass back to the processing of the block 76 .
- the blocks 78 - 80 can be performed only were the command is at the slot level. For example, the user has voiced commands to access an Email application and to perform a new message task but then pauses. At this point, the block 80 can prompt the user to enter information for a recipient slot, a subject line slot, and/or a message slot.
- the flowchart 60 of FIG. 2 is configured for use by beginner and expert users.
- an expert user can simply issue a continuous command that includes Application level, Action level, and/or Slot level audible inputs and the computing device will perform the desired command.
- a beginner user can issue separate commands for one or more of the Application, Action, and/or Slot levels and can be provided a prompt to assist, if needed.
- method 60 may include additional or fewer blocks, as needed or desired.
- the various blocks 62 - 76 may be combined into fewer blocks, divided into additional blocks, and/or removed based upon a desired implementation.
- each block 62 - 80 may represent a module, a segment, or a portion of program code, which includes one or more instructions executable by a processor for implementing specific logical functions or steps in the process.
- the program code may be stored on any type of computer readable medium or storage device including a disk or hard drive, for example.
- the computer readable medium may include non-transitory computer readable medium, for example, such as computer-readable media that stores data for short periods of time like register memory, processor cache and Random Access Memory (RAM).
- the computer readable medium may also include non-transitory media, such as secondary or persistent long term storage, like read only memory (ROM), optical or magnetic disks, compact-disc read only memory (CD-ROM), for example.
- the computer readable media may also be any other volatile or non-volatile storage systems.
- the computer readable medium may be considered a computer readable storage medium, for example, or a tangible storage device.
- one or more of the blocks 62 - 80 may represent circuitry that is wired to perform the specific logical functions of the method 60 .
- the user interface 100 includes a list 102 of various applications, which may have been identified during the block 62 of FIG. 2 .
- the applications include Calculator, Calendar, Camera, Messages, Compass, Email, Clock, Google+, Evernote, and Phone applications.
- One or more of these applications can be accessible by an audible input.
- all of the displayed applications are accessible by Application level audible inputs. More particularly, the block 64 of FIG.
- the block 66 displays the text in the list 102 .
- the text may correspond to a name of the application, although, in other examples, the text may include additional, fewer, and/or different text from the name of the application.
- a user has spoken the Application level audible input for the Email application, for example, which has been received by the block 68 .
- the blocks 70 - 74 have identified a plurality of actions, such as send or compose a new message, reply to a received message, read a most recent message, read all messages from a particular sender, show list of most recent messages (block 70 ), determined text of audible inputs for performing such actions, for example, New Message, Reply to, Read last, Read all from, and Show last (block 72 ), and displayed such text of audible inputs 104 (block 74 ).
- the text of Action level audible inputs 104 in FIG. 3B can be displayed automatically in response to receiving an Application level audible or manual user input to access the Email application. Further, in the present example, the displayed text 104 includes the actual text string that can be spoken to access each action.
- FIG. 3C the user has selected the new message action by speaking the audible input “New Message” or by manually selecting the New Message option from the user interface 100 of FIG. 3B . Consequently, the new message action is accessed and a new message user interface 106 is displayed, in accordance with the block 76 of FIG. 2 , for example.
- the prompt could involve a visual cue, such as highlighting the To slot 108 , and/or an audible cue to enter information in the To slot.
- the user has provided an input to select the To slot 108 and, in response, the user interface 100 has been updated to display a list of contacts 110 , which is shown in FIG. 3E , for example.
- the user can speak the name of an intended recipient or otherwise select the intended recipient.
- the user has selected Dennis, which populates the To slot 108 in the new message user interface 106 with an email address for Dennis, which is shown in FIG. 3F as implementis@example.com.
- another pause is identified by the block 78 and the block 80 prompts the user to enter an input for another slot, such as a Subject slot 112 or a Message slot 114 .
- the Message slot 114 is highlighted.
- the prompt could involve a visual cue, such as highlighting Message slot 114 and/or an audible cue to provide another slot input.
- the user provides an input, such as an audible input, to provide the message for the slot 114 , as shown in FIG. 3H .
- FIG. 3H further illustrates an example screenshot where the new message user interface 106 has been continuously and automatically updated in accordance with user inputs (for example, according to the block 76 of FIG. 3 ) and the identification of additional actions/slots (for example, according to the blocks 70 - 74 of FIG. 3 ). Consequently, in FIG. 3H , the user has provided inputs to the To slot 108 and the Message slot 114 , and is allowed to perform other actions represented by the text of audible inputs to Add more, Erase and start over, and Send, for example.
- FIG. 4 illustrates another example of a user providing a structured audible command in the context of a music player application.
- the structured audible command includes Application level, Action level, and Slot level inputs.
- various Action level and Slot level inputs can be provided and processed while the computing device is operating within a structured audible command.
- a spoken input “Music” is an Application level input that invokes a music application
- a spoken input “Play the Revolver album by The Beatles” is an Action level input that causes a particular album to be played.
- the user can provide an audible command as Slot level input to play the next track without providing audible commands to access the Application and Action levels.
- audible commands such as “Pause” and “Resume” could be used to control the play of tracks within the same Action level.
- Audible commands could also be used to obtain information within the Action level, such as “What is the next song?” Other types of audible commands could be used within a given Action level as well.
- spoken input can be received in a multi-level command format.
- the first level of the multi-level command format can be an Application level that relates to an application.
- the second level of the multi-level command format can be an Action level that relates to an action.
- the third level of the multi-level command format can be a Slot level that relates to slot input.
- a spoken utterance might include all three levels.
- a spoken utterance might include only the first level, or only the first and second levels. In that case, the one or more remaining levels of the multi-level command format could be included in one or more additional spoken utterances.
- a computing device could wait for a period of time to receive an additional spoken utterance that includes the second level. If no spoken utterance that includes the second level of the multi-level command format is received within the period of time, the computing device could provide a prompt. Similarly, if the spoken utterance and/or additional spoken utterance do not include the third level of the multi-level command format, then the computing device could wait for a period of time to receive an additional spoken utterance that includes the third level and, if not received within that period of time, provide a prompt.
- FIG. 5 illustrates an example method 200 for analyzing spoken input with reference to a multi-level command format.
- Block 202 involves receiving a spoken utterance (e.g., via a microphone of computing device 20 ) that includes at least a first level of a multi-level command format.
- the first level relates to an application.
- Block 204 involves causing speech recognition to be performed on the spoken utterance to obtain text corresponding to the spoken utterance.
- the speech recognition could be performed by the device that receives the spoken utterance (e.g., computing device 20 ), by a server in communication with the receiving device, or by a combination thereof.
- a computing device could transmit the spoken utterance to a server, the server could perform speech recognition on the spoken utterance to obtain text, and the computing device could receive the text from the server.
- Block 206 involves, based on the text, identifying the application. Identifying the application could involve use of predefined associations, in which each application is associated with one or more words or phrases that serve to identify the application. The application could then be identified by finding a word or phrase associated with that application in the text corresponding to the spoken utterance. For example, the word or phrase that identifies the application in the text could correspond to a word or phrase that is displayed on a display of a computing device as a visual cue, as discussed above for block 66 of FIG. 2 .
- Block 208 involves displaying a representation of the identified application.
- the representation of the identified application could be any display that is generated based on the identified application.
- the representation could be a user interface that is displayed on a display of a computing device.
- the representation could also include textual and/or graphical indications of actions that are available for that application (e.g., as shown in FIG. 3B ). Any of the indicated actions could be included in the spoken utterance or a subsequent spoken utterance as a second level of the multi-level command format.
- Block 210 involves determining whether the text includes a second level of the multi-level command format.
- the second level relates to an action. This determination could be made based on whether the text includes any words or phrases that are valid in the second level of the multi-command format, given the identified application as the first level of the multi-level command format.
- the valid words or phrases could be, for example, indicated in the representation of the identified application that is displayed at block 208 .
- Block 212 indicates a decision point in which a decision is made based on whether the text includes the second level of the multi-level command format. If the text includes the second level of the multi-level command format, method 200 proceeds to block 214 . If the text does not include the second level of the multi-level command format, method 200 proceeds to block 216 .
- Block 214 involves identifying the action and initiating the identified action. Identifying the action could involve use of predefined associations, in which each action is associated with a word or phrase. An action can then be identified by matching a word or phrase in the text that is valid for the second-level of the multi-level command format to a word or phrase that is associated with that action. Initiating the identified action could involve, for example, any activity by a computing device that causes the identified action to be performed on the computing device or on another device.
- Block 216 involves waiting (e.g., by computing device 20 ) for a first predetermined period of time to receive an additional utterance that includes the second level of the multi-level command format.
- Block 218 indicates a decision point in which a decision is made based on whether an additional utterance that includes the second level is received within the first predetermined period of time. If an additional utterance that includes the second level is received within the first predetermined period of time, method 200 proceeds to block 214 , as described above. If no additional utterance that includes the second level is received within the first predetermined period of time, then method 200 proceeds to block 220 .
- Block 220 involves providing (e.g., by computing device 20 ) at least one of an audible or visual action prompt.
- an additional spoken utterance that includes the second level could be received.
- method 200 could proceed to block 214 , as described above.
- the identified action could be initiated without any additional input. In other examples, however, the identified action could be initiated after populating one or more slots of the identified action with slot input.
- method 200 could further involve determining whether text obtained by speech recognition (e.g., speech recognition performed on either the spoken utterance received at block 202 or on an additional spoken utterance that is received within the waiting period of block 216 or after the prompt of block 220 ) includes a third level of the multi-level command format. The third level relates to slot input. If the text includes the third level, then one or more slots of the action identified in block 214 could be populated with the slot input provided as the third level. The action could then be initiated with the one or more slots populated with the slot input.
- speech recognition e.g., speech recognition performed on either the spoken utterance received at block 202 or on an additional spoken utterance that is received within the waiting period of block 216 or after the prompt of block 220 .
- the third level relates to slot input. If the text includes the third level, then one
- method 200 may involve waiting for a second predetermined period of time to receive an additional spoken utterance that includes the third level of the multi-level command format. If no spoken utterance that includes the third level is received within the second predetermined period of time, then an audible and/or visual prompt may be provided. In this way, an additional spoken utterance that includes the third level may be received within a waiting period or after a prompt. Slot input may be obtained from speech recognition that is performed on the additional spoken utterance and used to populate one or more slots of the identified action.
- Method 200 thus enables a voice interface to accept multiple levels of a multi-level command format in a single spoken utterance (e.g., a spoken utterance that identifies an application, an action, and slot input for the action) and to accept multiple levels of the multi-level command format in separate spoken utterances (e.g., an application in a first utterance, an action in a second utterance, and slot input in a third utterance).
- Each level of the multi-level command format may be obtained by speech recognition that is performed on one or more spoken utterances. The speech recognition could be performed by the device that receives the spoken utterance (e.g., computing device 20 ) and/or by a separate device (e.g., a server in communication with computing device 20 ).
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/621,018 US8452602B1 (en) | 2011-09-30 | 2012-09-15 | Structuring verbal commands to allow concatenation in a voice interface in a mobile device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161541992P | 2011-09-30 | 2011-09-30 | |
US13/621,018 US8452602B1 (en) | 2011-09-30 | 2012-09-15 | Structuring verbal commands to allow concatenation in a voice interface in a mobile device |
Publications (1)
Publication Number | Publication Date |
---|---|
US8452602B1 true US8452602B1 (en) | 2013-05-28 |
Family
ID=48445401
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/621,018 Active US8452602B1 (en) | 2011-09-30 | 2012-09-15 | Structuring verbal commands to allow concatenation in a voice interface in a mobile device |
Country Status (1)
Country | Link |
---|---|
US (1) | US8452602B1 (en) |
Cited By (92)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140095173A1 (en) * | 2012-10-01 | 2014-04-03 | Nuance Communications, Inc. | Systems and methods for providing a voice agent user interface |
US20140095167A1 (en) * | 2012-10-01 | 2014-04-03 | Nuance Communication, Inc. | Systems and methods for providing a voice agent user interface |
WO2015116151A1 (en) | 2014-01-31 | 2015-08-06 | Hewlett-Packard Development Company, L.P. | Voice input command |
WO2016017978A1 (en) * | 2014-07-31 | 2016-02-04 | Samsung Electronics Co., Ltd. | Device and method for performing functions |
US9342268B2 (en) | 2013-01-30 | 2016-05-17 | Google Inc. | Multi-level voice menu |
USD760754S1 (en) * | 2013-12-18 | 2016-07-05 | Apple Inc. | Display screen or portion thereof with graphical user interface |
USD762680S1 (en) * | 2013-12-18 | 2016-08-02 | Apple Inc. | Display screen or portion thereof with graphical user interface |
US9530410B1 (en) | 2013-04-09 | 2016-12-27 | Google Inc. | Multi-mode guard for voice commands |
CN106471570A (en) * | 2014-05-30 | 2017-03-01 | 苹果公司 | Multi-command single-speech input method |
US20170372703A1 (en) * | 2016-06-27 | 2017-12-28 | Google Inc. | Asynchronous processing of user requests |
US9922639B1 (en) * | 2013-01-11 | 2018-03-20 | Amazon Technologies, Inc. | User feedback for speech interactions |
US9953646B2 (en) | 2014-09-02 | 2018-04-24 | Belleau Technologies | Method and system for dynamic speech recognition and tracking of prewritten script |
USD823890S1 (en) | 2013-06-09 | 2018-07-24 | Apple Inc. | Display screen or portion thereof with icon |
US20180285065A1 (en) * | 2017-03-28 | 2018-10-04 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
US10607607B2 (en) * | 2016-01-07 | 2020-03-31 | Sony Corporation | Control device, display device, method, and program |
TWI695312B (en) * | 2014-07-31 | 2020-06-01 | 南韓商三星電子股份有限公司 | Device and method for performing functions |
US10720160B2 (en) | 2018-06-01 | 2020-07-21 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10741181B2 (en) | 2017-05-09 | 2020-08-11 | Apple Inc. | User interface for correcting recognition errors |
US10930282B2 (en) | 2015-03-08 | 2021-02-23 | Apple Inc. | Competing devices responding to voice triggers |
US10978090B2 (en) | 2013-02-07 | 2021-04-13 | Apple Inc. | Voice trigger for a digital assistant |
US11009970B2 (en) | 2018-06-01 | 2021-05-18 | Apple Inc. | Attention aware virtual assistant dismissal |
US11010127B2 (en) | 2015-06-29 | 2021-05-18 | Apple Inc. | Virtual assistant for media playback |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US11070949B2 (en) | 2015-05-27 | 2021-07-20 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on an electronic device with a touch-sensitive display |
US11087759B2 (en) | 2015-03-08 | 2021-08-10 | Apple Inc. | Virtual assistant activation |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US11126400B2 (en) | 2015-09-08 | 2021-09-21 | Apple Inc. | Zero latency digital assistant |
US11133008B2 (en) | 2014-05-30 | 2021-09-28 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11152002B2 (en) | 2016-06-11 | 2021-10-19 | Apple Inc. | Application integration with a digital assistant |
US11169616B2 (en) | 2018-05-07 | 2021-11-09 | Apple Inc. | Raise to speak |
US11217251B2 (en) | 2019-05-06 | 2022-01-04 | Apple Inc. | Spoken notifications |
USD941359S1 (en) | 2017-07-10 | 2022-01-18 | Apple Inc. | Display screen or portion thereof with icon |
US11237797B2 (en) | 2019-05-31 | 2022-02-01 | Apple Inc. | User activity shortcut suggestions |
US11257504B2 (en) | 2014-05-30 | 2022-02-22 | Apple Inc. | Intelligent assistant for home automation |
US11269678B2 (en) | 2012-05-15 | 2022-03-08 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
USD946607S1 (en) | 2020-06-19 | 2022-03-22 | Apple Inc. | Display screen or portion thereof with animated graphical user interface |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11348582B2 (en) | 2008-10-02 | 2022-05-31 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11380310B2 (en) | 2017-05-12 | 2022-07-05 | Apple Inc. | Low-latency intelligent automated assistant |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11431642B2 (en) | 2018-06-01 | 2022-08-30 | Apple Inc. | Variable latency device coordination |
US11468282B2 (en) | 2015-05-15 | 2022-10-11 | Apple Inc. | Virtual assistant in a communication session |
US11467802B2 (en) | 2017-05-11 | 2022-10-11 | Apple Inc. | Maintaining privacy of personal information |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US11516537B2 (en) | 2014-06-30 | 2022-11-29 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11532306B2 (en) | 2017-05-16 | 2022-12-20 | Apple Inc. | Detecting a trigger of a digital assistant |
US11580990B2 (en) | 2017-05-12 | 2023-02-14 | Apple Inc. | User-specific acoustic models |
US11599331B2 (en) | 2017-05-11 | 2023-03-07 | Apple Inc. | Maintaining privacy of personal information |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11657813B2 (en) | 2019-05-31 | 2023-05-23 | Apple Inc. | Voice identification in digital assistant systems |
US11656884B2 (en) | 2017-01-09 | 2023-05-23 | Apple Inc. | Application integration with a digital assistant |
US11671920B2 (en) | 2007-04-03 | 2023-06-06 | Apple Inc. | Method and system for operating a multifunction portable electronic device using voice-activation |
US11675829B2 (en) | 2017-05-16 | 2023-06-13 | Apple Inc. | Intelligent automated assistant for media exploration |
US11696060B2 (en) | 2020-07-21 | 2023-07-04 | Apple Inc. | User identification using headphones |
US11710482B2 (en) | 2018-03-26 | 2023-07-25 | Apple Inc. | Natural assistant interaction |
US11727219B2 (en) | 2013-06-09 | 2023-08-15 | Apple Inc. | System and method for inferring user intent from speech inputs |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11765209B2 (en) | 2020-05-11 | 2023-09-19 | Apple Inc. | Digital assistant hardware abstraction |
US11790914B2 (en) | 2019-06-01 | 2023-10-17 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11798547B2 (en) | 2013-03-15 | 2023-10-24 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US11809483B2 (en) | 2015-09-08 | 2023-11-07 | Apple Inc. | Intelligent automated assistant for media search and playback |
US11809783B2 (en) | 2016-06-11 | 2023-11-07 | Apple Inc. | Intelligent device arbitration and control |
US11838734B2 (en) | 2020-07-20 | 2023-12-05 | Apple Inc. | Multi-device audio adjustment coordination |
US11854539B2 (en) | 2018-05-07 | 2023-12-26 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11853536B2 (en) | 2015-09-08 | 2023-12-26 | Apple Inc. | Intelligent automated assistant in a media environment |
US11853647B2 (en) | 2015-12-23 | 2023-12-26 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US11880645B2 (en) | 2022-06-15 | 2024-01-23 | T-Mobile Usa, Inc. | Generating encoded text based on spoken utterances using machine learning systems and methods |
US11886805B2 (en) | 2015-11-09 | 2024-01-30 | Apple Inc. | Unconventional virtual assistant interactions |
US11893992B2 (en) | 2018-09-28 | 2024-02-06 | Apple Inc. | Multi-modal inputs for voice commands |
US11914848B2 (en) | 2020-05-11 | 2024-02-27 | Apple Inc. | Providing relevant data items based on context |
US11928604B2 (en) | 2005-09-08 | 2024-03-12 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US12010262B2 (en) | 2013-08-06 | 2024-06-11 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US12014118B2 (en) | 2017-05-15 | 2024-06-18 | Apple Inc. | Multi-modal interfaces having selection disambiguation and text modification capability |
US12051413B2 (en) | 2015-09-30 | 2024-07-30 | Apple Inc. | Intelligent device identification |
US12067985B2 (en) | 2018-06-01 | 2024-08-20 | Apple Inc. | Virtual assistant operations in multi-device environments |
US12073147B2 (en) | 2013-06-09 | 2024-08-27 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US12087308B2 (en) | 2010-01-18 | 2024-09-10 | Apple Inc. | Intelligent automated assistant |
US12197817B2 (en) | 2016-06-11 | 2025-01-14 | Apple Inc. | Intelligent device arbitration and control |
US12223282B2 (en) | 2016-06-09 | 2025-02-11 | Apple Inc. | Intelligent automated assistant in a home environment |
US12254887B2 (en) | 2017-05-16 | 2025-03-18 | Apple Inc. | Far-field extension of digital assistant services for providing a notification of an event to a user |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6208972B1 (en) * | 1998-12-23 | 2001-03-27 | Richard Grant | Method for integrating computer processes with an interface controlled by voice actuated grammars |
US20060200348A1 (en) * | 2002-04-08 | 2006-09-07 | Knott Benjamin A | Method and system for voice recognition menu navigation |
US7426469B1 (en) * | 1998-12-23 | 2008-09-16 | Eastern Investments Llc | Speech enabled computing method |
US20090327979A1 (en) * | 2008-06-30 | 2009-12-31 | Nokia Corporation | User interface for a peripheral device |
US20110223893A1 (en) * | 2009-09-30 | 2011-09-15 | T-Mobile Usa, Inc. | Genius Button Secondary Commands |
-
2012
- 2012-09-15 US US13/621,018 patent/US8452602B1/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6208972B1 (en) * | 1998-12-23 | 2001-03-27 | Richard Grant | Method for integrating computer processes with an interface controlled by voice actuated grammars |
US7426469B1 (en) * | 1998-12-23 | 2008-09-16 | Eastern Investments Llc | Speech enabled computing method |
US20060200348A1 (en) * | 2002-04-08 | 2006-09-07 | Knott Benjamin A | Method and system for voice recognition menu navigation |
US20090327979A1 (en) * | 2008-06-30 | 2009-12-31 | Nokia Corporation | User interface for a peripheral device |
US20110223893A1 (en) * | 2009-09-30 | 2011-09-15 | T-Mobile Usa, Inc. | Genius Button Secondary Commands |
Cited By (178)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11928604B2 (en) | 2005-09-08 | 2024-03-12 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US11671920B2 (en) | 2007-04-03 | 2023-06-06 | Apple Inc. | Method and system for operating a multifunction portable electronic device using voice-activation |
US11979836B2 (en) | 2007-04-03 | 2024-05-07 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US11348582B2 (en) | 2008-10-02 | 2022-05-31 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US11900936B2 (en) | 2008-10-02 | 2024-02-13 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US12165635B2 (en) | 2010-01-18 | 2024-12-10 | Apple Inc. | Intelligent automated assistant |
US12087308B2 (en) | 2010-01-18 | 2024-09-10 | Apple Inc. | Intelligent automated assistant |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US11269678B2 (en) | 2012-05-15 | 2022-03-08 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US11321116B2 (en) | 2012-05-15 | 2022-05-03 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US20140095173A1 (en) * | 2012-10-01 | 2014-04-03 | Nuance Communications, Inc. | Systems and methods for providing a voice agent user interface |
US20140095167A1 (en) * | 2012-10-01 | 2014-04-03 | Nuance Communication, Inc. | Systems and methods for providing a voice agent user interface |
US10276157B2 (en) * | 2012-10-01 | 2019-04-30 | Nuance Communications, Inc. | Systems and methods for providing a voice agent user interface |
US10950220B1 (en) * | 2013-01-11 | 2021-03-16 | Amazon Technologies, Inc. | User feedback for speech interactions |
US9922639B1 (en) * | 2013-01-11 | 2018-03-20 | Amazon Technologies, Inc. | User feedback for speech interactions |
US12008983B1 (en) * | 2013-01-11 | 2024-06-11 | Amazon Technologies, Inc. | User feedback for speech interactions |
US11990119B1 (en) * | 2013-01-11 | 2024-05-21 | Amazon Technologies, Inc. | User feedback for speech interactions |
US10460719B1 (en) * | 2013-01-11 | 2019-10-29 | Amazon Technologies, Inc. | User feedback for speech interactions |
US9342268B2 (en) | 2013-01-30 | 2016-05-17 | Google Inc. | Multi-level voice menu |
US9721567B2 (en) | 2013-01-30 | 2017-08-01 | Google Inc. | Multi-level voice menu |
US10019993B2 (en) | 2013-01-30 | 2018-07-10 | Google Llc | Multi-level voice menu |
US10319382B2 (en) | 2013-01-30 | 2019-06-11 | Google Llc | Multi-level voice menu |
US11557310B2 (en) | 2013-02-07 | 2023-01-17 | Apple Inc. | Voice trigger for a digital assistant |
US12009007B2 (en) | 2013-02-07 | 2024-06-11 | Apple Inc. | Voice trigger for a digital assistant |
US10978090B2 (en) | 2013-02-07 | 2021-04-13 | Apple Inc. | Voice trigger for a digital assistant |
US11862186B2 (en) | 2013-02-07 | 2024-01-02 | Apple Inc. | Voice trigger for a digital assistant |
US11636869B2 (en) | 2013-02-07 | 2023-04-25 | Apple Inc. | Voice trigger for a digital assistant |
US12277954B2 (en) | 2013-02-07 | 2025-04-15 | Apple Inc. | Voice trigger for a digital assistant |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US11798547B2 (en) | 2013-03-15 | 2023-10-24 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US10891953B2 (en) | 2013-04-09 | 2021-01-12 | Google Llc | Multi-mode guard for voice commands |
US9530410B1 (en) | 2013-04-09 | 2016-12-27 | Google Inc. | Multi-mode guard for voice commands |
US12293762B2 (en) | 2013-04-09 | 2025-05-06 | Google Llc | Multi-mode guard for voice commands |
US10181324B2 (en) | 2013-04-09 | 2019-01-15 | Google Llc | Multi-mode guard for voice commands |
US12073147B2 (en) | 2013-06-09 | 2024-08-27 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
USD823890S1 (en) | 2013-06-09 | 2018-07-24 | Apple Inc. | Display screen or portion thereof with icon |
US11727219B2 (en) | 2013-06-09 | 2023-08-15 | Apple Inc. | System and method for inferring user intent from speech inputs |
US12010262B2 (en) | 2013-08-06 | 2024-06-11 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
USD781899S1 (en) | 2013-12-18 | 2017-03-21 | Apple Inc. | Display screen or portion thereof with graphical user interface |
USD762680S1 (en) * | 2013-12-18 | 2016-08-02 | Apple Inc. | Display screen or portion thereof with graphical user interface |
USD760754S1 (en) * | 2013-12-18 | 2016-07-05 | Apple Inc. | Display screen or portion thereof with graphical user interface |
US10978060B2 (en) | 2014-01-31 | 2021-04-13 | Hewlett-Packard Development Company, L.P. | Voice input command |
EP3100259A4 (en) * | 2014-01-31 | 2017-08-30 | Hewlett-Packard Development Company, L.P. | Voice input command |
WO2015116151A1 (en) | 2014-01-31 | 2015-08-06 | Hewlett-Packard Development Company, L.P. | Voice input command |
CN105934791B (en) * | 2014-01-31 | 2019-11-22 | 惠普发展公司,有限责任合伙企业 | Voice input order |
CN105934791A (en) * | 2014-01-31 | 2016-09-07 | 惠普发展公司,有限责任合伙企业 | Voice input command |
US12067990B2 (en) | 2014-05-30 | 2024-08-20 | Apple Inc. | Intelligent assistant for home automation |
CN106471570B (en) * | 2014-05-30 | 2019-10-01 | 苹果公司 | Multi-command single-speech input method |
CN106471570A (en) * | 2014-05-30 | 2017-03-01 | 苹果公司 | Multi-command single-speech input method |
US11699448B2 (en) | 2014-05-30 | 2023-07-11 | Apple Inc. | Intelligent assistant for home automation |
US12118999B2 (en) | 2014-05-30 | 2024-10-15 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US11133008B2 (en) | 2014-05-30 | 2021-09-28 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10497365B2 (en) | 2014-05-30 | 2019-12-03 | Apple Inc. | Multi-command single utterance input method |
US11670289B2 (en) | 2014-05-30 | 2023-06-06 | Apple Inc. | Multi-command single utterance input method |
US10878809B2 (en) | 2014-05-30 | 2020-12-29 | Apple Inc. | Multi-command single utterance input method |
US11810562B2 (en) | 2014-05-30 | 2023-11-07 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US11257504B2 (en) | 2014-05-30 | 2022-02-22 | Apple Inc. | Intelligent assistant for home automation |
US11516537B2 (en) | 2014-06-30 | 2022-11-29 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US12200297B2 (en) | 2014-06-30 | 2025-01-14 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US11838579B2 (en) | 2014-06-30 | 2023-12-05 | Apple Inc. | Intelligent automated assistant for TV user interactions |
TWI695312B (en) * | 2014-07-31 | 2020-06-01 | 南韓商三星電子股份有限公司 | Device and method for performing functions |
US10768892B2 (en) | 2014-07-31 | 2020-09-08 | Samsung Electronics Co., Ltd. | Device and method for performing functions |
US10127011B2 (en) | 2014-07-31 | 2018-11-13 | Samsung Electronics Co., Ltd. | Device and method for performing functions |
WO2016017978A1 (en) * | 2014-07-31 | 2016-02-04 | Samsung Electronics Co., Ltd. | Device and method for performing functions |
US11099812B2 (en) | 2014-07-31 | 2021-08-24 | Samsung Electronics Co., Ltd. | Device and method for performing functions |
US9953646B2 (en) | 2014-09-02 | 2018-04-24 | Belleau Technologies | Method and system for dynamic speech recognition and tracking of prewritten script |
US12236952B2 (en) | 2015-03-08 | 2025-02-25 | Apple Inc. | Virtual assistant activation |
US11087759B2 (en) | 2015-03-08 | 2021-08-10 | Apple Inc. | Virtual assistant activation |
US11842734B2 (en) | 2015-03-08 | 2023-12-12 | Apple Inc. | Virtual assistant activation |
US10930282B2 (en) | 2015-03-08 | 2021-02-23 | Apple Inc. | Competing devices responding to voice triggers |
US12001933B2 (en) | 2015-05-15 | 2024-06-04 | Apple Inc. | Virtual assistant in a communication session |
US11468282B2 (en) | 2015-05-15 | 2022-10-11 | Apple Inc. | Virtual assistant in a communication session |
US12154016B2 (en) | 2015-05-15 | 2024-11-26 | Apple Inc. | Virtual assistant in a communication session |
US11070949B2 (en) | 2015-05-27 | 2021-07-20 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on an electronic device with a touch-sensitive display |
US11010127B2 (en) | 2015-06-29 | 2021-05-18 | Apple Inc. | Virtual assistant for media playback |
US11947873B2 (en) | 2015-06-29 | 2024-04-02 | Apple Inc. | Virtual assistant for media playback |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US11809483B2 (en) | 2015-09-08 | 2023-11-07 | Apple Inc. | Intelligent automated assistant for media search and playback |
US11550542B2 (en) | 2015-09-08 | 2023-01-10 | Apple Inc. | Zero latency digital assistant |
US11126400B2 (en) | 2015-09-08 | 2021-09-21 | Apple Inc. | Zero latency digital assistant |
US11954405B2 (en) | 2015-09-08 | 2024-04-09 | Apple Inc. | Zero latency digital assistant |
US12204932B2 (en) | 2015-09-08 | 2025-01-21 | Apple Inc. | Distributed personal assistant |
US11853536B2 (en) | 2015-09-08 | 2023-12-26 | Apple Inc. | Intelligent automated assistant in a media environment |
US12051413B2 (en) | 2015-09-30 | 2024-07-30 | Apple Inc. | Intelligent device identification |
US11809886B2 (en) | 2015-11-06 | 2023-11-07 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11886805B2 (en) | 2015-11-09 | 2024-01-30 | Apple Inc. | Unconventional virtual assistant interactions |
US11853647B2 (en) | 2015-12-23 | 2023-12-26 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10607607B2 (en) * | 2016-01-07 | 2020-03-31 | Sony Corporation | Control device, display device, method, and program |
US12223282B2 (en) | 2016-06-09 | 2025-02-11 | Apple Inc. | Intelligent automated assistant in a home environment |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US12175977B2 (en) | 2016-06-10 | 2024-12-24 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US11657820B2 (en) | 2016-06-10 | 2023-05-23 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US12293763B2 (en) | 2016-06-11 | 2025-05-06 | Apple Inc. | Application integration with a digital assistant |
US11152002B2 (en) | 2016-06-11 | 2021-10-19 | Apple Inc. | Application integration with a digital assistant |
US12197817B2 (en) | 2016-06-11 | 2025-01-14 | Apple Inc. | Intelligent device arbitration and control |
US11809783B2 (en) | 2016-06-11 | 2023-11-07 | Apple Inc. | Intelligent device arbitration and control |
US11749275B2 (en) | 2016-06-11 | 2023-09-05 | Apple Inc. | Application integration with a digital assistant |
US11302333B2 (en) * | 2016-06-27 | 2022-04-12 | Google Llc | Asynchronous processing of user requests |
US20170372703A1 (en) * | 2016-06-27 | 2017-12-28 | Google Inc. | Asynchronous processing of user requests |
US10339934B2 (en) * | 2016-06-27 | 2019-07-02 | Google Llc | Asynchronous processing of user requests |
US20190333519A1 (en) * | 2016-06-27 | 2019-10-31 | Google Llc | Asynchronous processing of user requests |
US10777204B2 (en) * | 2016-06-27 | 2020-09-15 | Google Llc | Asynchronous processing of user requests |
CN109196532A (en) * | 2016-06-27 | 2019-01-11 | 谷歌有限责任公司 | Asynchronous processing of digital assistant requests |
US11656884B2 (en) | 2017-01-09 | 2023-05-23 | Apple Inc. | Application integration with a digital assistant |
US12260234B2 (en) | 2017-01-09 | 2025-03-25 | Apple Inc. | Application integration with a digital assistant |
US20180285065A1 (en) * | 2017-03-28 | 2018-10-04 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
US11385861B2 (en) | 2017-03-28 | 2022-07-12 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
US10489111B2 (en) * | 2017-03-28 | 2019-11-26 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
US11372619B2 (en) | 2017-03-28 | 2022-06-28 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
US10741181B2 (en) | 2017-05-09 | 2020-08-11 | Apple Inc. | User interface for correcting recognition errors |
US11467802B2 (en) | 2017-05-11 | 2022-10-11 | Apple Inc. | Maintaining privacy of personal information |
US11599331B2 (en) | 2017-05-11 | 2023-03-07 | Apple Inc. | Maintaining privacy of personal information |
US11837237B2 (en) | 2017-05-12 | 2023-12-05 | Apple Inc. | User-specific acoustic models |
US11538469B2 (en) | 2017-05-12 | 2022-12-27 | Apple Inc. | Low-latency intelligent automated assistant |
US11580990B2 (en) | 2017-05-12 | 2023-02-14 | Apple Inc. | User-specific acoustic models |
US11862151B2 (en) | 2017-05-12 | 2024-01-02 | Apple Inc. | Low-latency intelligent automated assistant |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US11380310B2 (en) | 2017-05-12 | 2022-07-05 | Apple Inc. | Low-latency intelligent automated assistant |
US12014118B2 (en) | 2017-05-15 | 2024-06-18 | Apple Inc. | Multi-modal interfaces having selection disambiguation and text modification capability |
US11675829B2 (en) | 2017-05-16 | 2023-06-13 | Apple Inc. | Intelligent automated assistant for media exploration |
US11532306B2 (en) | 2017-05-16 | 2022-12-20 | Apple Inc. | Detecting a trigger of a digital assistant |
US12254887B2 (en) | 2017-05-16 | 2025-03-18 | Apple Inc. | Far-field extension of digital assistant services for providing a notification of an event to a user |
US12026197B2 (en) | 2017-05-16 | 2024-07-02 | Apple Inc. | Intelligent automated assistant for media exploration |
USD941359S1 (en) | 2017-07-10 | 2022-01-18 | Apple Inc. | Display screen or portion thereof with icon |
US11710482B2 (en) | 2018-03-26 | 2023-07-25 | Apple Inc. | Natural assistant interaction |
US12211502B2 (en) | 2018-03-26 | 2025-01-28 | Apple Inc. | Natural assistant interaction |
US11487364B2 (en) | 2018-05-07 | 2022-11-01 | Apple Inc. | Raise to speak |
US11900923B2 (en) | 2018-05-07 | 2024-02-13 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11854539B2 (en) | 2018-05-07 | 2023-12-26 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11907436B2 (en) | 2018-05-07 | 2024-02-20 | Apple Inc. | Raise to speak |
US11169616B2 (en) | 2018-05-07 | 2021-11-09 | Apple Inc. | Raise to speak |
US10984798B2 (en) | 2018-06-01 | 2021-04-20 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US12067985B2 (en) | 2018-06-01 | 2024-08-20 | Apple Inc. | Virtual assistant operations in multi-device environments |
US10720160B2 (en) | 2018-06-01 | 2020-07-21 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11009970B2 (en) | 2018-06-01 | 2021-05-18 | Apple Inc. | Attention aware virtual assistant dismissal |
US11360577B2 (en) | 2018-06-01 | 2022-06-14 | Apple Inc. | Attention aware virtual assistant dismissal |
US11431642B2 (en) | 2018-06-01 | 2022-08-30 | Apple Inc. | Variable latency device coordination |
US11630525B2 (en) | 2018-06-01 | 2023-04-18 | Apple Inc. | Attention aware virtual assistant dismissal |
US12061752B2 (en) | 2018-06-01 | 2024-08-13 | Apple Inc. | Attention aware virtual assistant dismissal |
US12080287B2 (en) | 2018-06-01 | 2024-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11893992B2 (en) | 2018-09-28 | 2024-02-06 | Apple Inc. | Multi-modal inputs for voice commands |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11783815B2 (en) | 2019-03-18 | 2023-10-10 | Apple Inc. | Multimodality in digital assistant systems |
US12136419B2 (en) | 2019-03-18 | 2024-11-05 | Apple Inc. | Multimodality in digital assistant systems |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11675491B2 (en) | 2019-05-06 | 2023-06-13 | Apple Inc. | User configurable task triggers |
US12216894B2 (en) | 2019-05-06 | 2025-02-04 | Apple Inc. | User configurable task triggers |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11705130B2 (en) | 2019-05-06 | 2023-07-18 | Apple Inc. | Spoken notifications |
US12154571B2 (en) | 2019-05-06 | 2024-11-26 | Apple Inc. | Spoken notifications |
US11217251B2 (en) | 2019-05-06 | 2022-01-04 | Apple Inc. | Spoken notifications |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11888791B2 (en) | 2019-05-21 | 2024-01-30 | Apple Inc. | Providing message response suggestions |
US11237797B2 (en) | 2019-05-31 | 2022-02-01 | Apple Inc. | User activity shortcut suggestions |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11657813B2 (en) | 2019-05-31 | 2023-05-23 | Apple Inc. | Voice identification in digital assistant systems |
US11360739B2 (en) | 2019-05-31 | 2022-06-14 | Apple Inc. | User activity shortcut suggestions |
US11790914B2 (en) | 2019-06-01 | 2023-10-17 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US12197712B2 (en) | 2020-05-11 | 2025-01-14 | Apple Inc. | Providing relevant data items based on context |
US11914848B2 (en) | 2020-05-11 | 2024-02-27 | Apple Inc. | Providing relevant data items based on context |
US11924254B2 (en) | 2020-05-11 | 2024-03-05 | Apple Inc. | Digital assistant hardware abstraction |
US11765209B2 (en) | 2020-05-11 | 2023-09-19 | Apple Inc. | Digital assistant hardware abstraction |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
USD946607S1 (en) | 2020-06-19 | 2022-03-22 | Apple Inc. | Display screen or portion thereof with animated graphical user interface |
US11838734B2 (en) | 2020-07-20 | 2023-12-05 | Apple Inc. | Multi-device audio adjustment coordination |
US12219314B2 (en) | 2020-07-21 | 2025-02-04 | Apple Inc. | User identification using headphones |
US11750962B2 (en) | 2020-07-21 | 2023-09-05 | Apple Inc. | User identification using headphones |
US11696060B2 (en) | 2020-07-21 | 2023-07-04 | Apple Inc. | User identification using headphones |
US12248748B2 (en) | 2022-06-15 | 2025-03-11 | T-Mobile Usa, Inc. | Generating encoded text based on spoken utterances using machine learning systems and methods |
US11880645B2 (en) | 2022-06-15 | 2024-01-23 | T-Mobile Usa, Inc. | Generating encoded text based on spoken utterances using machine learning systems and methods |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8452602B1 (en) | Structuring verbal commands to allow concatenation in a voice interface in a mobile device | |
US11037566B2 (en) | Word-level correction of speech input | |
US10489112B1 (en) | Method for user training of information dialogue system | |
CN108255290B (en) | Modal learning on mobile devices | |
US11790891B2 (en) | Wake word selection assistance architectures and methods | |
KR101703911B1 (en) | Visual confirmation for a recognized voice-initiated action | |
KR102115541B1 (en) | Speech re-recognition using external data sources | |
TW200900967A (en) | Multi-mode input method editor | |
TW201606750A (en) | Speech recognition using a foreign word grammar | |
US20210166687A1 (en) | Terminal device, server and controlling method thereof | |
US20200410991A1 (en) | System and method for predictive speech to text | |
CN116206601A (en) | Ordering method and device based on voice recognition, storage medium and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GOOGLE INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BRINGERT, BJORN ERIK;PIETRYKA, PAWEL;HODGSON, PETER JOHN;AND OTHERS;SIGNING DATES FROM 20120824 TO 20120914;REEL/FRAME:028971/0165 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: GOOGLE LLC, CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:GOOGLE INC.;REEL/FRAME:044101/0299 Effective date: 20170929 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |