US20050288936A1 - Multi-context conversational environment system and method - Google Patents

Multi-context conversational environment system and method Download PDF

Info

Publication number
US20050288936A1
US20050288936A1 US11/204,887 US20488705A US2005288936A1 US 20050288936 A1 US20050288936 A1 US 20050288936A1 US 20488705 A US20488705 A US 20488705A US 2005288936 A1 US2005288936 A1 US 2005288936A1
Authority
US
United States
Prior art keywords
application
speech
speech input
caller
subject
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/204,887
Inventor
Senis Busayapongchai
Pichet Chintrakulchai
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nuance Communications Inc
Original Assignee
BellSouth Intellectual Property Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BellSouth Intellectual Property Corp filed Critical BellSouth Intellectual Property Corp
Priority to US11/204,887 priority Critical patent/US20050288936A1/en
Publication of US20050288936A1 publication Critical patent/US20050288936A1/en
Assigned to AT&T INTELLECTUAL PROPERTY I, L.P. reassignment AT&T INTELLECTUAL PROPERTY I, L.P. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AT&T DELAWARE INTELLECTUAL PROPERTY, INC.
Assigned to BELLSOUTH INTELLECTUAL PROPERTY CORPORATION reassignment BELLSOUTH INTELLECTUAL PROPERTY CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BUSAYAPONGCHAI, SENIS, CHINTRAKULCHAI, PICHET
Assigned to AT&T INTELLECTUAL PROPERTY, INC. reassignment AT&T INTELLECTUAL PROPERTY, INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: BELLSOUTH INTELLECTUAL PROPERTY CORPORATION
Assigned to AT&T BLS INTELLECTUAL PROPERTY, INC. reassignment AT&T BLS INTELLECTUAL PROPERTY, INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: AT&T INTELLECTUAL PROPERTY, INC.
Assigned to AT&T DELAWARE INTELLECTUAL PROPERTY, INC. reassignment AT&T DELAWARE INTELLECTUAL PROPERTY, INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: AT&T BLS INTELLECTUAL PROPERTY, INC.
Assigned to AT&T ALEX HOLDINGS, LLC reassignment AT&T ALEX HOLDINGS, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AT&T INTELLECTUAL PROPERTY I, L.P.
Assigned to NUANCE COMMUNICATIONS, INC. reassignment NUANCE COMMUNICATIONS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AT&T ALEX HOLDINGS, LLC
Assigned to AT&T INTELLECTUAL PROPERTY I, L.P. reassignment AT&T INTELLECTUAL PROPERTY I, L.P. CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE/ASSIGNOR PREVIOUSLY RECORDED ON REEL 034467 FRAME 0853. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT. Assignors: AT&T DELAWARE INTELLECTUAL PROPERTY, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/487Arrangements for providing information services, e.g. recorded voice services or time announcements
    • H04M3/493Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
    • H04M3/4936Speech interaction details
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/226Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics
    • G10L2015/228Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics of application context

Definitions

  • the present invention relates in general to speech recognition and, more particularly, to interactive speech applications for use in automated telephone assistance systems.
  • Computer-based interactive speech applications are widely used in telephone systems to answer incoming calls as well as to perform various tasks including gathering information from callers, providing information to callers, and connecting callers with appropriate parties.
  • the call flow is activated when the system receives an incoming call, and begins with a system-generated greeting followed by a list of available options.
  • the application waits for a response from the caller and processes the response when received. The application must be able to recognize what the caller said to determine what should happen next.
  • the present invention is directed to providing a multi-contextual speech application environment wherein several applications can be activated at the same time, thereby enabling the user to switch between applications at will while maintaining each application context.
  • a control manager starts up applications, monitors speech for specific “control words” or “control phrases”, and switches control among applications.
  • Each application is independent of other applications, interacting with the control manager but not directly interacting with other applications, thereby leading to ease of development and maintenance because the first application does not need to be aware of the existence of other active applications.
  • FIG. 1 shows a block diagram of a general speech recognition system
  • FIG. 2 shows a more detailed block diagram of the system of FIG. 1 ;
  • FIGS. 3 and 4 depict portions of an exemplary call flow as is well known in the art
  • FIG. 5 shows a block diagram of an exemplary speech recognition system in accordance with the present invention
  • FIG. 6 shows a more detailed block diagram of an exemplary speech recognition system in accordance with the present invention.
  • FIG. 7 shows a flow diagram of an exemplary method of speech recognition in accordance with the present invention.
  • a typical speech recognition system generally comprises a device such as a microphone or telephone set 101 that converts a spoken utterance into an electric signal and transmits the signal to a speech recognition unit 100 .
  • the speech recognition unit 100 can be split into two functional blocks: a processing unit 102 and a search unit 104 .
  • the processing unit 102 is an acoustic processor that performs the segmentation, the normalization, and the parameterization of the input signal waveform. In some cases, especially for connected word speech, this stage may also include a feature extraction operation.
  • the search unit 104 includes a speech recognition dictionary that is scored in order to find possible matches to the spoken utterance.
  • the processing unit 102 translates the incoming analog speech waveform into digital format. This can be done with the use of an A/D converter 110 , a spectrogram generator or any other suitable device or technique.
  • the input signal is then split into short segments called analysis frames whose typical duration ranges from about 5-20 ms. Further processing will be done relative to these frames.
  • the processing unit 102 further comprises a feature extractor 112 , which can comprise a normalizer and a parameterizer, and an endpoint detector 114 .
  • the normalizer adjusts the maximum signal amplitude of each analysis frame to a standard level in order to take into account variations in speech intensity, transmission losses and other physical effects such as distance from the microphone and recording level.
  • the parameterizer typically represents speech frames in terms of voicing decision, amplitude and fundamental frequency. A wide variety of parameters can be used in the parameterizer.
  • the endpoint detector 114 splits the input signal waveform into a starting point and an endpoint of the speech utterance. This stage uses algorithms whose purpose is to locate the boundaries between silence and speech. Many systems use the short term energy and the zero crossing rate as an indication of the beginning or end of a word. Moreover, typical endpoint detection units use many parameters including frame energy, frame voice labels and other statistical variance parameters derived from speech.
  • the search unit 104 scores or otherwise ranks all the words (also known as “orthographies”) in a speech recognition dictionary database 120 such as to be able to derive the orthography or orthographies which have the highest probability of matching the spoken utterance.
  • a comparator and scorer 122 compares the spoken utterance with the entries in the database 120 to determine a match based on closest score.
  • Typical algorithms that can be used include the fast score estimation and the graph search algorithms, as known by those skilled in the art.
  • the accept/reject stage 124 compares the score to a predetermined threshold to determine if a correct mapping was found or if no mapping could be determined. If the score is greater than or equal to the predetermined threshold, then it is determined that a correct mapping has been found and the recognized word is provided as output, generally to another part of the system which may use the output as a link to information. If the score is below the threshold, no match is found and the input spoken utterance is rejected.
  • FIGS. 3 and 4 depict the call flow of an illustrative interactive speech application 300 as is presently known in the art, for use by a caller to obtain information.
  • Application 300 may be executed by a voice processing unit in a telephone system.
  • the call flow is activated when the system receives an incoming call, and begins by outputting a greeting at step 310 , such as, “Welcome to Info by Voice for Atlanta. Please tell me the information category you want now or say ‘Help’ and I will explain your choices.”
  • the application outputs an audible speech signal to the caller by, for example, playing a pre-recorded prompt or using a speech generator such as text-to-speech converter to produce the message.
  • the application then lists available options to the caller at step 320 , by outputting a message such as, “Please listen carefully to the following eleven choices. Remember you can speak your request at any time: Restaurant Guides, Ticket Information, Stock Quotes, Weather Forecast, Business News, Sports, Daily Horoscopes, Lottery Results”.
  • the system waits for a response from the caller at step 330 .
  • the application determines one or more potential matches between the user's input and the available options, creating a set of hypotheses. If the caller says, for example, “Restaurants”, the application must be able to recognize what the caller said and determine whether the caller's spoken input corresponds to one of the choices listed in the options. The application may determine that there are a plurality of potential matches between the caller's spoken input and the list of possible choices.
  • the application asks the user to repeat his selection and processing returns to step 330 , to wait for the caller's response.
  • the application attempts to confirm the caller's response. Typically, the application will attempt to verify what the caller said, by asking, for example, “I think you said ‘Restaurant’. Is that correct?”
  • step 360 the application once again waits to receive a response from the caller. If the caller says “yes” and his response is understood by the application, processing continues at step 370 . If the caller says “no” the caller will be re-prompted for input (not shown). When the caller's confirmation, (“Yes”), the application takes the appropriate steps to initiate the correct process at step 370 , which in the given example is process 305 a. This process repeats for each acceptable recognition result.
  • processing continues at step 370 .
  • the application then prompts the caller for a response by outputting a message such as “Welcome to ‘Restaurant’. Do you want to search by restaurant name or by type of food?” at step 325 a.
  • the system then waits for a response from the caller, at step 330 a, and processes the response when received at step 340 a. If the caller says, for example, “Type of food” the application must be able to recognize what the caller said and determine what information to provide.
  • the application again attempts to verify what the caller said, at step 350 a, by asking another question requiring a response, such as “I think you said ‘Search by type of food’.
  • the speech application will then say, for example, “Tell me the type of food you want. You may also include the area of town or landmark you want.” After this information is received from the caller (e.g., “Chinese restaurants in Midtown”) at step 370 a, the application lists the Chinese restaurants in Midtown at step 380 a.
  • processing for the first application terminates, and the call thread is lost. That is, if a caller who has navigated to “Chinese restaurants in Midtown” switches to “Weather” and then returns to “Restaurants”, the caller will have to start again at the top, “Do you want to search by name of restaurant or by type of food?” rather than picking up at listing the Chinese restaurants in Midtown.
  • the present invention is directed to systems and methods for providing a multi-contextual speech application environment wherein several applications can be open at the same time, enabling a caller to switch between applications at will.
  • a control manager interfaces with the caller, starts up applications, switches control among applications, alternately suspends and re-activates open applications, and maintains the context of suspended applications.
  • Each application may be independent of other applications. Alternately the context may be transferred across applications. For example if a user who asked for “Chinese in Atlanta”, then switched to weather, the system may ask the user if the user in interested in Atlanta weather. Additionally, the chance of recognizing “Atlanta” compared to other cities may be increased.
  • the present invention enables a caller to open multiple threads at the same time. For example, the present invention would enable a caller to navigate to “Chinese restaurants in Midtown,” and then switch to “Weather,” and then return to “Restaurants” and once again be at the point where the Chinese restaurants in Midtown would be listed.
  • any number of applications can be open at any time, limited by storage and other processing considerations. Such an environment helps simplify the application design because the application designer concentrates only on the designer's application. System users, however, may mix and match the functionalities of multiple applications according to the user's needs.
  • FIG. 5 illustrates an exemplary environment in which the present invention may be deployed.
  • a caller connects to a communications network 504 such as the publicly switched telephone network (PSTN) though a communications device 502 such as a telephone to receive audible information from a database of information stored on and processed by a server 506 .
  • a communications network 504 such as the publicly switched telephone network (PSTN)
  • PSTN publicly switched telephone network
  • a communications device 502 such as a telephone to receive audible information from a database of information stored on and processed by a server 506 .
  • Components of a typical server computer may include, but are not limited to, a processing unit, a system memory, a system bus that couples various system components including the system memory to the processing unit, storage media and various input and output devices.
  • the communications device 502 may include a subscriber telephone set that is connected via a telephone line (e.g., POTS, or similar) to a telephone system (e.g., communications network 504 ) including at least one central office switch, at least one service control point (SCP), and a service node (SN).
  • a PSTN e.g., PSTN
  • AIN Advanced Intelligent Network
  • the SN typically can include voice and dual tone multi-frequency (DTMF) signal recognition devices and voice synthesis devices which can be used in accordance with the present invention.
  • DTMF dual tone multi-frequency
  • FIG. 6 shows a more detailed block diagram of an exemplary speech recognition system in accordance with the present invention.
  • the conversational operating environment 600 acts as a mediating agent between a caller's speech input 630 and the application software 620 a, 620 b, 620 c through speech input and voice prompting output.
  • Each open application 620 a, 620 b, 620 c is assigned to a task agent, analogous to a window in the graphical world.
  • Control manager 604 starts up applications, assigns a newly started application 620 a, 620 b, 620 c to a task agent 610 a, 610 b, 610 c, switches control among task agents 610 a, 610 b, 610 c, maintains the context for each application in context table 602 , and may transfers context across applications.
  • the caller can call up control manager 604 at any time from any application 620 a, 620 b, 620 c by speaking one of a list of control words or phrases.
  • the caller interacts with the operating environment 600 and at any given time operating environment 600 directs the caller's speech input 630 , when appropriate, to the currently active task agent 610 a, 610 b, 610 c.
  • the current environment states for a particular caller can persistently be saved and restored across sessions in context table 602 for each caller.
  • a caller may request to be reminded of the last context for an application 620 a, 620 b, 620 c.
  • Control manager 604 will output to the caller the context stored in context table 602 for the application 620 a, 620 b, 620 c.
  • FIG. 7 shows a flow diagram of an exemplary method of speech recognition in accordance with the present invention.
  • a caller initiates a call to an information service in accordance with the present invention.
  • a welcome message is desirably played to the caller and the caller thereafter is prompted for the name of a desired service application (for example, a subject area such as “Weather”, “Stock Quotes”, etc.).
  • the caller indicates the desired application or subject area by speaking into the telephone handset, for example.
  • the speech input is then processed/searched to determine the desired application. If the speech input does not correspond to an application, the user is prompted to repeat the speech input, for example, or the call may be forwarded to a human operator for assistance.
  • the control manager initiates a task agent for the appropriate (user selected) application, as described above.
  • the task agent then activates the application, at step 720 .
  • the application prompts the caller for input through the control manager.
  • the input is used by the application to narrow or limit the search (e.g., after choosing “Restaurants”, the input may be for the type of restaurant or location, for example).
  • the input is received by the control manager and processed/searched to determine whether the input corresponds to available selections. If the speech input does not correspond to an application, the user is prompted to repeat the speech input, for example, or the call may be forwarded to a human operator for assistance.
  • the control manager monitors the input received and analyzes the input for the presence of command words, such as those that correspond to starting another application or subject area (e.g., “Main Menu”, “Lottery Results”, “Stock Quotes”, etc.).
  • command words such as those that correspond to starting another application or subject area (e.g., “Main Menu”, “Lottery Results”, “Stock Quotes”, etc.).
  • the input is passed to the current speech application at step 736 , which uses it in the current search or thread.
  • the control manager saves the context of the currently open application in the context table at step 740 .
  • the command word is compared to a list of words indicating that the caller wishes to terminate the call (e.g., “Goodbye”, “End”, etc.). If it is determined that the caller wishes to terminate the call, the system exits at step 756 . If termination is not indicated, the command word is compared to a list of possible applications to determine the next application to be initiated, at step 748 .
  • the control manager suspends the first application at step 752 and initiates a task agent for the second application, with processing continuing at step 716 .
  • At least one indicator is stored that indicates the current processing step of the initial application. In this manner, the caller can return to the application at the point where it was suspended when the second application was opened.
  • the at least one indicator comprises a series of indicators that indicate a processing path of the first application.
  • a computer-generated representation of the series of indicators is stored (for future use or analysis) that indicates the processing path of the first application.
  • a caller initiates a call to a telephone information service in accordance with the present invention in which the information provided includes weather, airline reservations and hotel reservations.
  • the caller may indicate “Airline reservation” when prompted for his choice of application, and navigate to the point where he will receive information concerning the cost of a flight from Atlanta to Philadelphia on August 4. At this point the caller may say “Hotel” to indicate that he wishes to receive information concerning hotel reservations. After determining the cost or availability of hotel reservations in Philadelphia, the caller could indicate his desire to return to the “Airline Reservation” application by saying “Airline” and would be returned to the point at which he suspended the “Airline Reservation” application to inquire about hotel reservations.
  • the caller could indicate his desire to return to the “Airline Reservation” application by saying “Airline” and would be returned to the point at which he suspended the “Airline Reservation” application to inquire about hotel reservations.
  • each application can provide the audible information to the caller in an associated voice, so that different voices are used for different applications. In this manner, a caller will hear distinguishable voices providing information for the various applications.

Abstract

An interactive speech-activated information retrieval application for use in automated telephone systems includes a control manager that interfaces between the caller's speech input and applications and enables several applications to be open at the same time. The control manager continually monitors for control words, enabling the user to switch between applications at will. When a user switches to another application, the control manager suspends the first application and stores its context, enabling the user to later return to the application at the point where the application was previously suspended.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation application and claims priority to U.S. patent application Ser. No. 09/870,202, filed May 30, 2001, entitled “Multi-Context Conversational Environment System And Method”, which is hereby incorporated by reference in its entirety.
  • FIELD OF THE INVENTION
  • The present invention relates in general to speech recognition and, more particularly, to interactive speech applications for use in automated telephone assistance systems.
  • BACKGROUND OF THE INVENTION
  • Computer-based interactive speech applications are widely used in telephone systems to answer incoming calls as well as to perform various tasks including gathering information from callers, providing information to callers, and connecting callers with appropriate parties. Typically, the call flow is activated when the system receives an incoming call, and begins with a system-generated greeting followed by a list of available options. The application waits for a response from the caller and processes the response when received. The application must be able to recognize what the caller said to determine what should happen next. It does this by converting the user's speech into a text sentence of distinct words (speech recognition), breaking down the recognized sentence grammatically and then systematically representing its meaning (language understanding), obtaining targeted data based on that meaning's representation from an appropriate online source (information retrieval), building a text sentence that presents the retrieved data in the user's preferred language (language generation), and converting that text sentence into computer-generated speech (speech synthesis).
  • Typically, interactive speech applications today are “single-threaded.” If a user accesses an interactive speech-activated system to receive information, he would select one of the available options, complete the process and then return to the main menu to select a second option, if desired. If a user abandons the process before completion, the navigation through the system so far is lost and the user must start again at the top of the call flow. It would be helpful if a user were able to select an option, go partway through the call flow, decide to access another thread before proceeding with the first thread, suspend the first thread, select a second thread, and thereafter be able to return to the first thread at the point he left it. Hence a need exists for a multi-contextual speech application that overcomes the drawbacks of the prior art.
  • SUMMARY OF THE INVENTION
  • The present invention is directed to providing a multi-contextual speech application environment wherein several applications can be activated at the same time, thereby enabling the user to switch between applications at will while maintaining each application context. A control manager starts up applications, monitors speech for specific “control words” or “control phrases”, and switches control among applications. Each application is independent of other applications, interacting with the control manager but not directly interacting with other applications, thereby leading to ease of development and maintenance because the first application does not need to be aware of the existence of other active applications.
  • The foregoing and other aspects of the present invention will become apparent from the following detailed description of the invention when considered in conjunction with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 shows a block diagram of a general speech recognition system;
  • FIG. 2 shows a more detailed block diagram of the system of FIG. 1;
  • FIGS. 3 and 4 depict portions of an exemplary call flow as is well known in the art;
  • FIG. 5 shows a block diagram of an exemplary speech recognition system in accordance with the present invention;
  • FIG. 6 shows a more detailed block diagram of an exemplary speech recognition system in accordance with the present invention; and
  • FIG. 7 shows a flow diagram of an exemplary method of speech recognition in accordance with the present invention.
  • DESCRIPTION OF EXEMPLARY EMBODIMENTS AND BEST MODE
  • Overview
  • Speech recognition systems have been developed in many parts of the world and, although it is difficult to describe a standard recognition system architecture, some characteristics are shared between many of them. A typical speech recognition system, of the type depicted in FIG. 1, generally comprises a device such as a microphone or telephone set 101 that converts a spoken utterance into an electric signal and transmits the signal to a speech recognition unit 100. The speech recognition unit 100 can be split into two functional blocks: a processing unit 102 and a search unit 104. The processing unit 102 is an acoustic processor that performs the segmentation, the normalization, and the parameterization of the input signal waveform. In some cases, especially for connected word speech, this stage may also include a feature extraction operation. The search unit 104 includes a speech recognition dictionary that is scored in order to find possible matches to the spoken utterance.
  • More specifically, the processing unit 102, illustrated in greater detail in FIG. 2, translates the incoming analog speech waveform into digital format. This can be done with the use of an A/D converter 110, a spectrogram generator or any other suitable device or technique. The input signal is then split into short segments called analysis frames whose typical duration ranges from about 5-20 ms. Further processing will be done relative to these frames. The processing unit 102 further comprises a feature extractor 112, which can comprise a normalizer and a parameterizer, and an endpoint detector 114.
  • The normalizer adjusts the maximum signal amplitude of each analysis frame to a standard level in order to take into account variations in speech intensity, transmission losses and other physical effects such as distance from the microphone and recording level. The parameterizer typically represents speech frames in terms of voicing decision, amplitude and fundamental frequency. A wide variety of parameters can be used in the parameterizer.
  • The endpoint detector 114 splits the input signal waveform into a starting point and an endpoint of the speech utterance. This stage uses algorithms whose purpose is to locate the boundaries between silence and speech. Many systems use the short term energy and the zero crossing rate as an indication of the beginning or end of a word. Moreover, typical endpoint detection units use many parameters including frame energy, frame voice labels and other statistical variance parameters derived from speech.
  • The search unit 104, shown in more detail in FIG. 2, scores or otherwise ranks all the words (also known as “orthographies”) in a speech recognition dictionary database 120 such as to be able to derive the orthography or orthographies which have the highest probability of matching the spoken utterance. A comparator and scorer 122 compares the spoken utterance with the entries in the database 120 to determine a match based on closest score. Typical algorithms that can be used include the fast score estimation and the graph search algorithms, as known by those skilled in the art.
  • The accept/reject stage 124 compares the score to a predetermined threshold to determine if a correct mapping was found or if no mapping could be determined. If the score is greater than or equal to the predetermined threshold, then it is determined that a correct mapping has been found and the recognized word is provided as output, generally to another part of the system which may use the output as a link to information. If the score is below the threshold, no match is found and the input spoken utterance is rejected.
  • FIGS. 3 and 4 depict the call flow of an illustrative interactive speech application 300 as is presently known in the art, for use by a caller to obtain information. Application 300 may be executed by a voice processing unit in a telephone system. The call flow is activated when the system receives an incoming call, and begins by outputting a greeting at step 310, such as, “Welcome to Info by Voice for Atlanta. Please tell me the information category you want now or say ‘Help’ and I will explain your choices.” In this example, the application outputs an audible speech signal to the caller by, for example, playing a pre-recorded prompt or using a speech generator such as text-to-speech converter to produce the message.
  • If the user responds, “Help” at step 315, the application then lists available options to the caller at step 320, by outputting a message such as, “Please listen carefully to the following eleven choices. Remember you can speak your request at any time: Restaurant Guides, Ticket Information, Stock Quotes, Weather Forecast, Business News, Sports, Daily Horoscopes, Lottery Results”.
  • If the caller does not say “Help” at step 315 or after the options are listed at step 320, the system waits for a response from the caller at step 330. At step 340 the application determines one or more potential matches between the user's input and the available options, creating a set of hypotheses. If the caller says, for example, “Restaurants”, the application must be able to recognize what the caller said and determine whether the caller's spoken input corresponds to one of the choices listed in the options. The application may determine that there are a plurality of potential matches between the caller's spoken input and the list of possible choices. At step 342, if there are a plurality of potential matches, the application asks the user to repeat his selection and processing returns to step 330, to wait for the caller's response.
  • If the set of hypotheses contains only one potential match but the score of the recognized string is lower than a predetermined confirmation threshold, at step 350 the application attempts to confirm the caller's response. Typically, the application will attempt to verify what the caller said, by asking, for example, “I think you said ‘Restaurant’. Is that correct?”
  • At step 360, the application once again waits to receive a response from the caller. If the caller says “yes” and his response is understood by the application, processing continues at step 370. If the caller says “no” the caller will be re-prompted for input (not shown). When the caller's confirmation, (“Yes”), the application takes the appropriate steps to initiate the correct process at step 370, which in the given example is process 305 a. This process repeats for each acceptable recognition result.
  • If the set of hypotheses contains only one potential match and the score of the recognized string is higher than the predetermined confirmation threshold, processing continues at step 370.
  • Referring now to FIG. 4, the application then prompts the caller for a response by outputting a message such as “Welcome to ‘Restaurant’. Do you want to search by restaurant name or by type of food?” at step 325 a. The system then waits for a response from the caller, at step 330 a, and processes the response when received at step 340 a. If the caller says, for example, “Type of food” the application must be able to recognize what the caller said and determine what information to provide. The application again attempts to verify what the caller said, at step 350 a, by asking another question requiring a response, such as “I think you said ‘Search by type of food’. Is that correct?” If the caller responds, at step 360 a, “Yes”, the speech application will then say, for example, “Tell me the type of food you want. You may also include the area of town or landmark you want.” After this information is received from the caller (e.g., “Chinese restaurants in Midtown”) at step 370 a, the application lists the Chinese restaurants in Midtown at step 380 a.
  • These types of conversational environments offered by current speech recognition applications allow a caller to interact using only a one-to-one dialog mode. For example, in order to obtain information about another topic, such as “Weather” in the example above and shown in process 305 b, the caller must exit the “Restaurant” application and return to the main menu by saying “Main Menu”, or go to the beginning of the weather service by saying “Weather”. Processing then continues similar to that described above with respect to process 305 a, but with a different set of prompts and information.
  • Additionally, whenever a request to switch to another topic option is received, processing for the first application terminates, and the call thread is lost. That is, if a caller who has navigated to “Chinese restaurants in Midtown” switches to “Weather” and then returns to “Restaurants”, the caller will have to start again at the top, “Do you want to search by name of restaurant or by type of food?” rather than picking up at listing the Chinese restaurants in Midtown.
  • Multi-Context Conversational Environment
  • The present invention is directed to systems and methods for providing a multi-contextual speech application environment wherein several applications can be open at the same time, enabling a caller to switch between applications at will. A control manager interfaces with the caller, starts up applications, switches control among applications, alternately suspends and re-activates open applications, and maintains the context of suspended applications. Each application may be independent of other applications. Alternately the context may be transferred across applications. For example if a user who asked for “Chinese in Atlanta”, then switched to weather, the system may ask the user if the user in interested in Atlanta weather. Additionally, the chance of recognizing “Atlanta” compared to other cities may be increased.
  • The present invention enables a caller to open multiple threads at the same time. For example, the present invention would enable a caller to navigate to “Chinese restaurants in Midtown,” and then switch to “Weather,” and then return to “Restaurants” and once again be at the point where the Chinese restaurants in Midtown would be listed. Moreover, any number of applications can be open at any time, limited by storage and other processing considerations. Such an environment helps simplify the application design because the application designer concentrates only on the designer's application. System users, however, may mix and match the functionalities of multiple applications according to the user's needs.
  • FIG. 5 illustrates an exemplary environment in which the present invention may be deployed. As can be seen, a caller connects to a communications network 504 such as the publicly switched telephone network (PSTN) though a communications device 502 such as a telephone to receive audible information from a database of information stored on and processed by a server 506. Components of a typical server computer may include, but are not limited to, a processing unit, a system memory, a system bus that couples various system components including the system memory to the processing unit, storage media and various input and output devices.
  • The communications device 502 may include a subscriber telephone set that is connected via a telephone line (e.g., POTS, or similar) to a telephone system (e.g., communications network 504) including at least one central office switch, at least one service control point (SCP), and a service node (SN). This exemplary environment is a PSTN. It is contemplated that a portion of the PSTN may be part of the Advanced Intelligent Network (AIN) of a typical local exchange carrier. For further information regarding the referenced PSTN and AIN aspects thereof, the interested reader is referred to the patent to Weisser, U.S. Pat. No. 5,430,719, which is incorporated herein by reference. The SN typically can include voice and dual tone multi-frequency (DTMF) signal recognition devices and voice synthesis devices which can be used in accordance with the present invention.
  • FIG. 6 shows a more detailed block diagram of an exemplary speech recognition system in accordance with the present invention. The conversational operating environment 600 acts as a mediating agent between a caller's speech input 630 and the application software 620 a, 620 b, 620 c through speech input and voice prompting output. Each open application 620 a, 620 b, 620 c is assigned to a task agent, analogous to a window in the graphical world. Control manager 604 starts up applications, assigns a newly started application 620 a, 620 b, 620 c to a task agent 610 a, 610 b, 610 c, switches control among task agents 610 a, 610 b, 610 c, maintains the context for each application in context table 602, and may transfers context across applications. The caller can call up control manager 604 at any time from any application 620 a, 620 b, 620 c by speaking one of a list of control words or phrases. The caller interacts with the operating environment 600 and at any given time operating environment 600 directs the caller's speech input 630, when appropriate, to the currently active task agent 610 a, 610 b, 610 c.
  • Because almost all of the processing for current over-the-telephone speech applications occur on the server side (the client is merely a telephone and has little or no processing power), the current environment states for a particular caller can persistently be saved and restored across sessions in context table 602 for each caller. According to one aspect of the invention, a caller may request to be reminded of the last context for an application 620 a, 620 b, 620 c. Control manager 604 will output to the caller the context stored in context table 602 for the application 620 a, 620 b, 620 c. Although three applications are described with respect to FIG. 6, it is contemplated that any number of applications can be implemented.
  • FIG. 7 shows a flow diagram of an exemplary method of speech recognition in accordance with the present invention. At step 704, a caller initiates a call to an information service in accordance with the present invention. At step 708, a welcome message is desirably played to the caller and the caller thereafter is prompted for the name of a desired service application (for example, a subject area such as “Weather”, “Stock Quotes”, etc.). At step 712, the caller indicates the desired application or subject area by speaking into the telephone handset, for example. The speech input is then processed/searched to determine the desired application. If the speech input does not correspond to an application, the user is prompted to repeat the speech input, for example, or the call may be forwarded to a human operator for assistance.
  • At step 716, the control manager initiates a task agent for the appropriate (user selected) application, as described above. The task agent then activates the application, at step 720. At step 724, the application prompts the caller for input through the control manager. The input is used by the application to narrow or limit the search (e.g., after choosing “Restaurants”, the input may be for the type of restaurant or location, for example). At step 728, the input is received by the control manager and processed/searched to determine whether the input corresponds to available selections. If the speech input does not correspond to an application, the user is prompted to repeat the speech input, for example, or the call may be forwarded to a human operator for assistance.
  • The control manager monitors the input received and analyzes the input for the presence of command words, such as those that correspond to starting another application or subject area (e.g., “Main Menu”, “Lottery Results”, “Stock Quotes”, etc.). At step 732, if no command words are found, the input is passed to the current speech application at step 736, which uses it in the current search or thread.
  • On the other hand, if at step 732 it is determined that the caller's input does contain a command word or words, the control manager saves the context of the currently open application in the context table at step 740. At step 744, the command word is compared to a list of words indicating that the caller wishes to terminate the call (e.g., “Goodbye”, “End”, etc.). If it is determined that the caller wishes to terminate the call, the system exits at step 756. If termination is not indicated, the command word is compared to a list of possible applications to determine the next application to be initiated, at step 748. If a valid service is found, the control manager suspends the first application at step 752 and initiates a task agent for the second application, with processing continuing at step 716. At least one indicator is stored that indicates the current processing step of the initial application. In this manner, the caller can return to the application at the point where it was suspended when the second application was opened. Preferably, the at least one indicator comprises a series of indicators that indicate a processing path of the first application. Moreover, desirably, a computer-generated representation of the series of indicators is stored (for future use or analysis) that indicates the processing path of the first application.
  • As an example, assume that a caller initiates a call to a telephone information service in accordance with the present invention in which the information provided includes weather, airline reservations and hotel reservations. The caller may indicate “Airline reservation” when prompted for his choice of application, and navigate to the point where he will receive information concerning the cost of a flight from Atlanta to Philadelphia on August 4. At this point the caller may say “Hotel” to indicate that he wishes to receive information concerning hotel reservations. After determining the cost or availability of hotel reservations in Philadelphia, the caller could indicate his desire to return to the “Airline Reservation” application by saying “Airline” and would be returned to the point at which he suspended the “Airline Reservation” application to inquire about hotel reservations. Although in this example only two contexts are illustrated, it should be understood that any number of threads or applications could be active at the same time, limited only by the capacity of the context table.
  • It is contemplated that each application can provide the audible information to the caller in an associated voice, so that different voices are used for different applications. In this manner, a caller will hear distinguishable voices providing information for the various applications.
  • Although illustrated and described herein with reference to certain specific embodiments, the present invention is nevertheless not intended to be limited to the details shown. Rather, various modifications may be made in the details within the scope and range of equivalents of the claims and without departing from the invention.

Claims (20)

1. A method of speech recognition processing that provides audible information over a communications device comprising:
receiving a first speech input indicative of a first subject area;
initiating a first subject application associated with said first subject area;
receiving a second speech input indicative of a second subject area; and
storing at least one indicator indicating a current processing step of said first subject application.
2. The method according to claim 1, further comprising initiating a second subject application associated with said second subject area.
3. The method according to claim 1, further comprising initiating a task agent for said first subject application.
4. The method according to claim 1, further comprising monitoring said first speech input for at least one word indicative of said second subject area.
5. The method according to claim 1, further comprising suspending said first subject application after receiving said second speech input.
6. The method according to claim 5, further comprising:
receiving a further speech input, and
re-activating said first subject application responsive to the further speech input.
7. The method according to claim 1, wherein said storing at least one indicator further comprises storing a series of indicators that indicate a processing path of said first application.
8. The method according to claim 7, further comprising outputting a computer-generated representation of said stored series of indicators that indicates said processing path of said first application.
9. The method according to claim 1, further comprising outputting information associated with said first application in a first voice.
10. The method according to claim 9, further comprising outputting information associated with said second application in a second voice, said second voice being distinguishable from said first voice.
11. The method according to claim 1, further comprising synthesizing speech for outputting speech from said first application.
12. A speech recognition system comprising:
a speech recognition module that processes speech input and translates said speech input into computer-readable input;
a control manager comprising:
a module that interfaces between said speech input and at least one of a plurality of application programs;
a module that initiates processing of a first application program; and
a module that monitors said speech input for a request to initiate a second application program; and
a speech synthesizing module for providing output information from said plurality of application programs.
13. The system according to claim 12, further comprising a context table for maintaining a context for each of said plurality of application programs.
14. The system according to claim 12, further comprising a plurality of task agents, each task agent associated with one of said plurality of application programs.
15. The system according to claim 14, wherein the control manager is adapted to assign said application programs to said task agents, and switch control among said task agents.
16. The system according to claim 12, wherein the request comprises a control word.
17. The system according to claim 12, wherein the control manager is adapted to suspend said first application program, and initiate processing of said second application program, responsive to the request.
18. The system according to claim 17, wherein the control manager is adapted to re-activate said first application program responsive to a further request.
19. The system according to claim 12, wherein the control manager is adapted to store at least one indicator indicative of a current processing step of at least one of said plurality of application programs.
20. A computer-readable medium for storing computer-executable instructions for performing the method of claim 1.
US11/204,887 2001-05-30 2005-08-15 Multi-context conversational environment system and method Abandoned US20050288936A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/204,887 US20050288936A1 (en) 2001-05-30 2005-08-15 Multi-context conversational environment system and method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US09/870,202 US6944594B2 (en) 2001-05-30 2001-05-30 Multi-context conversational environment system and method
US11/204,887 US20050288936A1 (en) 2001-05-30 2005-08-15 Multi-context conversational environment system and method

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09/870,202 Continuation US6944594B2 (en) 2001-05-30 2001-05-30 Multi-context conversational environment system and method

Publications (1)

Publication Number Publication Date
US20050288936A1 true US20050288936A1 (en) 2005-12-29

Family

ID=25354946

Family Applications (2)

Application Number Title Priority Date Filing Date
US09/870,202 Expired - Lifetime US6944594B2 (en) 2001-05-30 2001-05-30 Multi-context conversational environment system and method
US11/204,887 Abandoned US20050288936A1 (en) 2001-05-30 2005-08-15 Multi-context conversational environment system and method

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US09/870,202 Expired - Lifetime US6944594B2 (en) 2001-05-30 2001-05-30 Multi-context conversational environment system and method

Country Status (2)

Country Link
US (2) US6944594B2 (en)
WO (1) WO2002097795A1 (en)

Cited By (158)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070208568A1 (en) * 2006-03-04 2007-09-06 At&T Corp. Menu Hierarchy Skipping Dialog For Directed Dialog Speech Recognition
US20090248412A1 (en) * 2008-03-27 2009-10-01 Fujitsu Limited Association apparatus, association method, and recording medium
US20090326930A1 (en) * 2006-07-12 2009-12-31 Panasonic Corporation Speech decoding apparatus and speech encoding apparatus
KR20120137440A (en) * 2010-01-18 2012-12-20 애플 인크. Maintaining context information between user interactions with a voice assistant
US8670985B2 (en) 2010-01-13 2014-03-11 Apple Inc. Devices and methods for identifying a prompt corresponding to a voice input in a sequence of prompts
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US8682649B2 (en) 2009-11-12 2014-03-25 Apple Inc. Sentiment prediction from textual data
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US8688446B2 (en) 2008-02-22 2014-04-01 Apple Inc. Providing text input using speech data and non-speech data
US8706472B2 (en) 2011-08-11 2014-04-22 Apple Inc. Method for disambiguating multiple readings in language conversion
US8713021B2 (en) 2010-07-07 2014-04-29 Apple Inc. Unsupervised document clustering using latent semantic density analysis
US8712776B2 (en) 2008-09-29 2014-04-29 Apple Inc. Systems and methods for selective text to speech synthesis
US8718047B2 (en) 2001-10-22 2014-05-06 Apple Inc. Text to speech conversion of text messages from mobile communication devices
US8719006B2 (en) 2010-08-27 2014-05-06 Apple Inc. Combined statistical and rule-based part-of-speech tagging for text-to-speech synthesis
US8719014B2 (en) 2010-09-27 2014-05-06 Apple Inc. Electronic device with text error correction based on voice recognition data
US8751238B2 (en) 2009-03-09 2014-06-10 Apple Inc. Systems and methods for determining the language to use for speech generated by a text to speech engine
US8762156B2 (en) 2011-09-28 2014-06-24 Apple Inc. Speech recognition repair using contextual information
US8768702B2 (en) 2008-09-05 2014-07-01 Apple Inc. Multi-tiered voice feedback in an electronic device
US8775442B2 (en) 2012-05-15 2014-07-08 Apple Inc. Semantic search using a single-source semantic model
US8781836B2 (en) 2011-02-22 2014-07-15 Apple Inc. Hearing assistance system for providing consistent human speech
US8812294B2 (en) 2011-06-21 2014-08-19 Apple Inc. Translating phrases from one language into another using an order-based set of declarative rules
US8862252B2 (en) 2009-01-30 2014-10-14 Apple Inc. Audio user interface for displayless electronic device
US8898568B2 (en) 2008-09-09 2014-11-25 Apple Inc. Audio user interface
US8935167B2 (en) 2012-09-25 2015-01-13 Apple Inc. Exemplar-based latent perceptual modeling for automatic speech recognition
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US8977584B2 (en) 2010-01-25 2015-03-10 Newvaluexchange Global Ai Llp Apparatuses, methods and systems for a digital conversation management platform
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US9053089B2 (en) 2007-10-02 2015-06-09 Apple Inc. Part-of-speech tagging using latent analogy
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
AU2013205584B2 (en) * 2010-01-18 2016-03-24 Apple Inc. Maintaining context information between user interactions with a voice assistant
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US9311043B2 (en) 2010-01-13 2016-04-12 Apple Inc. Adaptive audio feedback system and method
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9733821B2 (en) 2013-03-14 2017-08-15 Apple Inc. Voice control to diagnose inadvertent activation of accessibility features
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US20170308389A1 (en) * 2014-09-30 2017-10-26 Nuance Communications, Inc. Methods And Apparatus For Module Arbitration
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US20180012601A1 (en) * 2013-11-18 2018-01-11 Amazon Technologies, Inc. Dialog management with multiple applications
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9916831B2 (en) 2014-05-30 2018-03-13 Yandex Europe Ag System and method for handling a spoken user request
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9946706B2 (en) 2008-06-07 2018-04-17 Apple Inc. Automatic language identification for dynamic text processing
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US9977779B2 (en) 2013-03-14 2018-05-22 Apple Inc. Automatic supplementation of word correction dictionaries
US10019994B2 (en) 2012-06-08 2018-07-10 Apple Inc. Systems and methods for recognizing textual identifiers within a plurality of words
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10078487B2 (en) 2013-03-15 2018-09-18 Apple Inc. Context-sensitive handling of interruptions
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10255566B2 (en) 2011-06-03 2019-04-09 Apple Inc. Generating and processing task items that represent tasks to perform
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US10356243B2 (en) 2015-06-05 2019-07-16 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US10410637B2 (en) 2017-05-12 2019-09-10 Apple Inc. User-specific acoustic models
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US10482874B2 (en) 2017-05-15 2019-11-19 Apple Inc. Hierarchical belief states for digital assistants
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10515147B2 (en) 2010-12-22 2019-12-24 Apple Inc. Using statistical language models for contextual lookup
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10540976B2 (en) 2009-06-05 2020-01-21 Apple Inc. Contextual voice commands
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US10572476B2 (en) 2013-03-14 2020-02-25 Apple Inc. Refining a search based on schedule items
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10642574B2 (en) 2013-03-14 2020-05-05 Apple Inc. Device, method, and graphical user interface for outputting captions
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10672399B2 (en) 2011-06-03 2020-06-02 Apple Inc. Switching between text data and audio data based on a mapping
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10748529B1 (en) 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10755703B2 (en) 2017-05-11 2020-08-25 Apple Inc. Offline personal assistant
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US11151899B2 (en) 2013-03-15 2021-10-19 Apple Inc. User training by intelligent digital assistant
US11217255B2 (en) 2017-05-16 2022-01-04 Apple Inc. Far-field extension for digital assistant services
RU2785950C2 (en) * 2010-01-18 2022-12-15 Эппл Инк. Maintenance of contextual information between user interactions with voice assistant
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification

Families Citing this family (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7167832B2 (en) * 2001-10-15 2007-01-23 At&T Corp. Method for dialog management
DE10156945A1 (en) * 2001-11-20 2003-05-28 Philips Corp Intellectual Pty Method for operation of speech recognition dialogue system for provision of telephone information, etc., whereby navigation is improved as a user can jump straight to a point in a dialogue by giving a marker name
US8126713B2 (en) * 2002-04-11 2012-02-28 Shengyang Huang Conversation control system and conversation control method
US20030216915A1 (en) * 2002-05-15 2003-11-20 Jianlei Xie Voice command and voice recognition for hand-held devices
US7398209B2 (en) 2002-06-03 2008-07-08 Voicebox Technologies, Inc. Systems and methods for responding to natural language speech utterance
US7693720B2 (en) * 2002-07-15 2010-04-06 Voicebox Technologies, Inc. Mobile systems and methods for responding to natural language speech utterance
CN100346625C (en) * 2002-12-27 2007-10-31 联想(北京)有限公司 Telephone voice interactive system and its realizing method
JP2004287016A (en) * 2003-03-20 2004-10-14 Sony Corp Apparatus and method for speech interaction, and robot apparatus
US20040236679A1 (en) * 2003-05-20 2004-11-25 Anderson David J. Method and system for performing automated prepaid account renewal
US7280651B2 (en) * 2003-05-20 2007-10-09 At&T Bls Intellectual Property, Inc. Method and system for performing automated telemarketing
US7742580B2 (en) * 2004-02-05 2010-06-22 Avaya, Inc. Methods and apparatus for context and experience sensitive prompting in voice applications
US20050229185A1 (en) * 2004-02-20 2005-10-13 Stoops Daniel S Method and system for navigating applications
US7925512B2 (en) * 2004-05-19 2011-04-12 Nuance Communications, Inc. Method, system, and apparatus for a voice markup language interpreter and voice browser
US7640160B2 (en) 2005-08-05 2009-12-29 Voicebox Technologies, Inc. Systems and methods for responding to natural language speech utterance
US7620549B2 (en) 2005-08-10 2009-11-17 Voicebox Technologies, Inc. System and method of supporting adaptive misrecognition in conversational speech
US8126716B2 (en) * 2005-08-19 2012-02-28 Nuance Communications, Inc. Method and system for collecting audio prompts in a dynamically generated voice application
US7949529B2 (en) 2005-08-29 2011-05-24 Voicebox Technologies, Inc. Mobile systems and methods of supporting natural language human-machine interactions
WO2007027989A2 (en) 2005-08-31 2007-03-08 Voicebox Technologies, Inc. Dynamic speech sharpening
JP4849663B2 (en) * 2005-10-21 2012-01-11 株式会社ユニバーサルエンターテインメント Conversation control device
JP4846336B2 (en) * 2005-10-21 2011-12-28 株式会社ユニバーサルエンターテインメント Conversation control device
JP4849662B2 (en) * 2005-10-21 2012-01-11 株式会社ユニバーサルエンターテインメント Conversation control device
US8332218B2 (en) 2006-06-13 2012-12-11 Nuance Communications, Inc. Context-based grammars for automated speech recognition
US7746986B2 (en) * 2006-06-15 2010-06-29 Verizon Data Services Llc Methods and systems for a sign language graphical interpreter
AU2007297597B2 (en) * 2006-09-21 2013-02-21 Kyorin Pharmaceuticals Co., Ltd. Serine hydrolase inhibitors
US8073681B2 (en) 2006-10-16 2011-12-06 Voicebox Technologies, Inc. System and method for a cooperative conversational voice user interface
WO2008092473A1 (en) * 2007-01-31 2008-08-07 Telecom Italia S.P.A. Customizable method and system for emotional recognition
US7818176B2 (en) 2007-02-06 2010-10-19 Voicebox Technologies, Inc. System and method for selecting and presenting advertisements based on natural language processing of voice-based input
US8140335B2 (en) 2007-12-11 2012-03-20 Voicebox Technologies, Inc. System and method for providing a natural language voice user interface in an integrated voice navigation services environment
US8589161B2 (en) 2008-05-27 2013-11-19 Voicebox Technologies, Inc. System and method for an integrated, multi-modal, multi-device natural language voice services environment
US9305548B2 (en) 2008-05-27 2016-04-05 Voicebox Technologies Corporation System and method for an integrated, multi-modal, multi-device natural language voice services environment
US20110246187A1 (en) * 2008-12-16 2011-10-06 Koninklijke Philips Electronics N.V. Speech signal processing
US8326637B2 (en) 2009-02-20 2012-12-04 Voicebox Technologies, Inc. System and method for processing multi-modal device interactions in a natural language voice services environment
US9502025B2 (en) 2009-11-10 2016-11-22 Voicebox Technologies Corporation System and method for providing a natural language content dedication service
US9171541B2 (en) 2009-11-10 2015-10-27 Voicebox Technologies Corporation System and method for hybrid processing in a natural language voice services environment
US10026394B1 (en) * 2012-08-31 2018-07-17 Amazon Technologies, Inc. Managing dialogs on a speech recognition platform
US9112807B2 (en) * 2012-12-26 2015-08-18 Alcatel Lucent System and method for managing network information
US9431008B2 (en) 2013-05-29 2016-08-30 Nuance Communications, Inc. Multiple parallel dialogs in smart phone applications
US9626703B2 (en) 2014-09-16 2017-04-18 Voicebox Technologies Corporation Voice commerce
US9898459B2 (en) 2014-09-16 2018-02-20 Voicebox Technologies Corporation Integration of domain information into state transitions of a finite state transducer for natural language processing
EP3207467A4 (en) 2014-10-15 2018-05-23 VoiceBox Technologies Corporation System and method for providing follow-up responses to prior natural language inputs of a user
US10431214B2 (en) 2014-11-26 2019-10-01 Voicebox Technologies Corporation System and method of determining a domain and/or an action related to a natural language input
US10614799B2 (en) 2014-11-26 2020-04-07 Voicebox Technologies Corporation System and method of providing intent predictions for an utterance prior to a system detection of an end of the utterance
CN108475507A (en) * 2016-01-28 2018-08-31 索尼公司 Information processing equipment, information processing method and program
US10331784B2 (en) 2016-07-29 2019-06-25 Voicebox Technologies Corporation System and method of disambiguating natural language processing requests
US10997359B2 (en) 2018-04-27 2021-05-04 International Business Machines Corporation Real-time cognitive modifying a mark-up language document
CN113573132B (en) * 2021-07-23 2023-08-11 深圳康佳电子科技有限公司 Multi-application screen spelling method and device based on voice realization and storage medium

Citations (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5179585A (en) * 1991-01-16 1993-01-12 Octel Communications Corporation Integrated voice messaging/voice response system
US5430719A (en) * 1993-06-28 1995-07-04 Bellsouth Corporation Mediation of open advanced intelligent network interface by shared execution environment
US5544232A (en) * 1992-11-17 1996-08-06 Rockwell International Corporation Call distributor with automatic preannouncement system and method
US5615296A (en) * 1993-11-12 1997-03-25 International Business Machines Corporation Continuous speech recognition and voice response system and method to enable conversational dialogues with microprocessors
US5632002A (en) * 1992-12-28 1997-05-20 Kabushiki Kaisha Toshiba Speech recognition interface system suitable for window systems and speech mail systems
US5651096A (en) * 1995-03-14 1997-07-22 Apple Computer, Inc. Merging of language models from two or more application programs for a speech recognition system
US5699534A (en) * 1994-04-15 1997-12-16 International Business Machines Corporation Multiple display pointers for computer graphical user interfaces
US5699486A (en) * 1993-11-24 1997-12-16 Canon Information Systems, Inc. System for speaking hypertext documents such as computerized help files
US5765130A (en) * 1996-05-21 1998-06-09 Applied Language Technologies, Inc. Method and apparatus for facilitating speech barge-in in connection with voice recognition systems
US5812977A (en) * 1996-08-13 1998-09-22 Applied Voice Recognition L.P. Voice control computer interface enabling implementation of common subroutines
US5897618A (en) * 1997-03-10 1999-04-27 International Business Machines Corporation Data processing system and method for switching between programs having a same title using a voice command
US5918213A (en) * 1995-12-22 1999-06-29 Mci Communications Corporation System and method for automated remote previewing and purchasing of music, video, software, and other multimedia products
US5949857A (en) * 1998-12-17 1999-09-07 International Business Machines Corporation Telephone DTMF signal accessible data processor with calculator program
US6081774A (en) * 1997-08-22 2000-06-27 Novell, Inc. Natural language information retrieval system and method
US6115711A (en) * 1989-09-28 2000-09-05 Sterling Software, Inc. Method and apparatus for generating transactions and a dialog flow manager
US6125347A (en) * 1993-09-29 2000-09-26 L&H Applications Usa, Inc. System for controlling multiple user application programs by spoken input
US6173266B1 (en) * 1997-05-06 2001-01-09 Speechworks International, Inc. System and method for developing interactive speech applications
US6182046B1 (en) * 1998-03-26 2001-01-30 International Business Machines Corp. Managing voice commands in speech applications
US6192339B1 (en) * 1998-11-04 2001-02-20 Intel Corporation Mechanism for managing multiple speech applications
US6208972B1 (en) * 1998-12-23 2001-03-27 Richard Grant Method for integrating computer processes with an interface controlled by voice actuated grammars
US6233559B1 (en) * 1998-04-01 2001-05-15 Motorola, Inc. Speech control of multiple applications using applets
US6324507B1 (en) * 1999-02-10 2001-11-27 International Business Machines Corp. Speech recognition enrollment for non-readers and displayless devices
US6505162B1 (en) * 1999-06-11 2003-01-07 Industrial Technology Research Institute Apparatus and method for portable dialogue management using a hierarchial task description table
US6510411B1 (en) * 1999-10-29 2003-01-21 Unisys Corporation Task oriented dialog model and manager
US6603836B1 (en) * 1996-11-28 2003-08-05 British Telecommunications Public Limited Company Interactive voice response apparatus capable of distinguishing between user's incoming voice and outgoing conditioned voice prompts
US6615172B1 (en) * 1999-11-12 2003-09-02 Phoenix Solutions, Inc. Intelligent query engine for processing voice based queries
US6711618B1 (en) * 1999-09-03 2004-03-23 Cisco Technology, Inc. Apparatus and method for providing server state and attribute management for voice enabled web applications
US6727830B2 (en) * 1999-01-05 2004-04-27 Microsoft Corporation Time based hardware button for application launch
US6757365B1 (en) * 2000-10-16 2004-06-29 Tellme Networks, Inc. Instant messaging via telephone interfaces
US6883015B1 (en) * 2000-03-30 2005-04-19 Cisco Technology, Inc. Apparatus and method for providing server state and attribute management for multiple-threaded voice enabled web applications
US6985865B1 (en) * 2001-09-26 2006-01-10 Sprint Spectrum L.P. Method and system for enhanced response to voice commands in a voice command platform

Patent Citations (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6115711A (en) * 1989-09-28 2000-09-05 Sterling Software, Inc. Method and apparatus for generating transactions and a dialog flow manager
US5179585A (en) * 1991-01-16 1993-01-12 Octel Communications Corporation Integrated voice messaging/voice response system
US5544232A (en) * 1992-11-17 1996-08-06 Rockwell International Corporation Call distributor with automatic preannouncement system and method
US5632002A (en) * 1992-12-28 1997-05-20 Kabushiki Kaisha Toshiba Speech recognition interface system suitable for window systems and speech mail systems
US5430719A (en) * 1993-06-28 1995-07-04 Bellsouth Corporation Mediation of open advanced intelligent network interface by shared execution environment
US6125347A (en) * 1993-09-29 2000-09-26 L&H Applications Usa, Inc. System for controlling multiple user application programs by spoken input
US5615296A (en) * 1993-11-12 1997-03-25 International Business Machines Corporation Continuous speech recognition and voice response system and method to enable conversational dialogues with microprocessors
US5699486A (en) * 1993-11-24 1997-12-16 Canon Information Systems, Inc. System for speaking hypertext documents such as computerized help files
US5699534A (en) * 1994-04-15 1997-12-16 International Business Machines Corporation Multiple display pointers for computer graphical user interfaces
US5651096A (en) * 1995-03-14 1997-07-22 Apple Computer, Inc. Merging of language models from two or more application programs for a speech recognition system
US5918213A (en) * 1995-12-22 1999-06-29 Mci Communications Corporation System and method for automated remote previewing and purchasing of music, video, software, and other multimedia products
US5765130A (en) * 1996-05-21 1998-06-09 Applied Language Technologies, Inc. Method and apparatus for facilitating speech barge-in in connection with voice recognition systems
US5812977A (en) * 1996-08-13 1998-09-22 Applied Voice Recognition L.P. Voice control computer interface enabling implementation of common subroutines
US6603836B1 (en) * 1996-11-28 2003-08-05 British Telecommunications Public Limited Company Interactive voice response apparatus capable of distinguishing between user's incoming voice and outgoing conditioned voice prompts
US5897618A (en) * 1997-03-10 1999-04-27 International Business Machines Corporation Data processing system and method for switching between programs having a same title using a voice command
US6173266B1 (en) * 1997-05-06 2001-01-09 Speechworks International, Inc. System and method for developing interactive speech applications
US6081774A (en) * 1997-08-22 2000-06-27 Novell, Inc. Natural language information retrieval system and method
US6182046B1 (en) * 1998-03-26 2001-01-30 International Business Machines Corp. Managing voice commands in speech applications
US6233559B1 (en) * 1998-04-01 2001-05-15 Motorola, Inc. Speech control of multiple applications using applets
US6192339B1 (en) * 1998-11-04 2001-02-20 Intel Corporation Mechanism for managing multiple speech applications
US5949857A (en) * 1998-12-17 1999-09-07 International Business Machines Corporation Telephone DTMF signal accessible data processor with calculator program
US6208972B1 (en) * 1998-12-23 2001-03-27 Richard Grant Method for integrating computer processes with an interface controlled by voice actuated grammars
US6727830B2 (en) * 1999-01-05 2004-04-27 Microsoft Corporation Time based hardware button for application launch
US6324507B1 (en) * 1999-02-10 2001-11-27 International Business Machines Corp. Speech recognition enrollment for non-readers and displayless devices
US6505162B1 (en) * 1999-06-11 2003-01-07 Industrial Technology Research Institute Apparatus and method for portable dialogue management using a hierarchial task description table
US6711618B1 (en) * 1999-09-03 2004-03-23 Cisco Technology, Inc. Apparatus and method for providing server state and attribute management for voice enabled web applications
US6510411B1 (en) * 1999-10-29 2003-01-21 Unisys Corporation Task oriented dialog model and manager
US6615172B1 (en) * 1999-11-12 2003-09-02 Phoenix Solutions, Inc. Intelligent query engine for processing voice based queries
US6883015B1 (en) * 2000-03-30 2005-04-19 Cisco Technology, Inc. Apparatus and method for providing server state and attribute management for multiple-threaded voice enabled web applications
US6757365B1 (en) * 2000-10-16 2004-06-29 Tellme Networks, Inc. Instant messaging via telephone interfaces
US6985865B1 (en) * 2001-09-26 2006-01-10 Sprint Spectrum L.P. Method and system for enhanced response to voice commands in a voice command platform

Cited By (236)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US8718047B2 (en) 2001-10-22 2014-05-06 Apple Inc. Text to speech conversion of text messages from mobile communication devices
US9501741B2 (en) 2005-09-08 2016-11-22 Apple Inc. Method and apparatus for building an intelligent automated assistant
US10318871B2 (en) 2005-09-08 2019-06-11 Apple Inc. Method and apparatus for building an intelligent automated assistant
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US8862477B2 (en) 2006-03-04 2014-10-14 At&T Intellectual Property Ii, L.P. Menu hierarchy skipping dialog for directed dialog speech recognition
US20070208568A1 (en) * 2006-03-04 2007-09-06 At&T Corp. Menu Hierarchy Skipping Dialog For Directed Dialog Speech Recognition
US8457973B2 (en) 2006-03-04 2013-06-04 AT&T Intellectual Propert II, L.P. Menu hierarchy skipping dialog for directed dialog speech recognition
US8812306B2 (en) * 2006-07-12 2014-08-19 Panasonic Intellectual Property Corporation Of America Speech decoding and encoding apparatus for lost frame concealment using predetermined number of waveform samples peripheral to the lost frame
US20090326930A1 (en) * 2006-07-12 2009-12-31 Panasonic Corporation Speech decoding apparatus and speech encoding apparatus
US9117447B2 (en) 2006-09-08 2015-08-25 Apple Inc. Using event alert text as input to an automated assistant
US8942986B2 (en) 2006-09-08 2015-01-27 Apple Inc. Determining user intent based on ontologies of domains
US8930191B2 (en) 2006-09-08 2015-01-06 Apple Inc. Paraphrasing of user requests and results by automated digital assistant
US10568032B2 (en) 2007-04-03 2020-02-18 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US9053089B2 (en) 2007-10-02 2015-06-09 Apple Inc. Part-of-speech tagging using latent analogy
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US10381016B2 (en) 2008-01-03 2019-08-13 Apple Inc. Methods and apparatus for altering audio output signals
US8688446B2 (en) 2008-02-22 2014-04-01 Apple Inc. Providing text input using speech data and non-speech data
US9361886B2 (en) 2008-02-22 2016-06-07 Apple Inc. Providing text input using speech data and non-speech data
US20090248412A1 (en) * 2008-03-27 2009-10-01 Fujitsu Limited Association apparatus, association method, and recording medium
US9865248B2 (en) 2008-04-05 2018-01-09 Apple Inc. Intelligent text-to-speech conversion
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US9946706B2 (en) 2008-06-07 2018-04-17 Apple Inc. Automatic language identification for dynamic text processing
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US10108612B2 (en) 2008-07-31 2018-10-23 Apple Inc. Mobile device having human language translation capability with positional feedback
US9691383B2 (en) 2008-09-05 2017-06-27 Apple Inc. Multi-tiered voice feedback in an electronic device
US8768702B2 (en) 2008-09-05 2014-07-01 Apple Inc. Multi-tiered voice feedback in an electronic device
US8898568B2 (en) 2008-09-09 2014-11-25 Apple Inc. Audio user interface
US8712776B2 (en) 2008-09-29 2014-04-29 Apple Inc. Systems and methods for selective text to speech synthesis
US8762469B2 (en) 2008-10-02 2014-06-24 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US8713119B2 (en) 2008-10-02 2014-04-29 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US11900936B2 (en) 2008-10-02 2024-02-13 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US11348582B2 (en) 2008-10-02 2022-05-31 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US10643611B2 (en) 2008-10-02 2020-05-05 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US8862252B2 (en) 2009-01-30 2014-10-14 Apple Inc. Audio user interface for displayless electronic device
US8751238B2 (en) 2009-03-09 2014-06-10 Apple Inc. Systems and methods for determining the language to use for speech generated by a text to speech engine
US10540976B2 (en) 2009-06-05 2020-01-21 Apple Inc. Contextual voice commands
US10795541B2 (en) 2009-06-05 2020-10-06 Apple Inc. Intelligent organization of tasks items
US10475446B2 (en) 2009-06-05 2019-11-12 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US11080012B2 (en) 2009-06-05 2021-08-03 Apple Inc. Interface for a virtual digital assistant
US10283110B2 (en) 2009-07-02 2019-05-07 Apple Inc. Methods and apparatuses for automatic speech recognition
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US8682649B2 (en) 2009-11-12 2014-03-25 Apple Inc. Sentiment prediction from textual data
US9311043B2 (en) 2010-01-13 2016-04-12 Apple Inc. Adaptive audio feedback system and method
US8670985B2 (en) 2010-01-13 2014-03-11 Apple Inc. Devices and methods for identifying a prompt corresponding to a voice input in a sequence of prompts
RU2785950C2 (en) * 2010-01-18 2022-12-15 Эппл Инк. Maintenance of contextual information between user interactions with voice assistant
US20130117022A1 (en) * 2010-01-18 2013-05-09 Apple Inc. Personalized Vocabulary for Digital Assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
AU2013205584B2 (en) * 2010-01-18 2016-03-24 Apple Inc. Maintaining context information between user interactions with a voice assistant
US11423886B2 (en) 2010-01-18 2022-08-23 Apple Inc. Task flow identification based on user intent
KR20120137440A (en) * 2010-01-18 2012-12-20 애플 인크. Maintaining context information between user interactions with a voice assistant
US20130110518A1 (en) * 2010-01-18 2013-05-02 Apple Inc. Active Input Elicitation by Intelligent Automated Assistant
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10706841B2 (en) 2010-01-18 2020-07-07 Apple Inc. Task flow identification based on user intent
US8903716B2 (en) * 2010-01-18 2014-12-02 Apple Inc. Personalized vocabulary for digital assistant
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10741185B2 (en) 2010-01-18 2020-08-11 Apple Inc. Intelligent automated assistant
US8731942B2 (en) * 2010-01-18 2014-05-20 Apple Inc. Maintaining context information between user interactions with a voice assistant
US8660849B2 (en) 2010-01-18 2014-02-25 Apple Inc. Prioritizing selection criteria by automated assistant
RU2653250C2 (en) * 2010-01-18 2018-05-07 Эппл Инк. Support of context information during interactions between user and voice assistant
US9548050B2 (en) 2010-01-18 2017-01-17 Apple Inc. Intelligent automated assistant
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
KR101588081B1 (en) * 2010-01-18 2016-01-25 애플 인크. Maintaining context information between user interactions with a voice assistant
US8670979B2 (en) * 2010-01-18 2014-03-11 Apple Inc. Active input elicitation by intelligent automated assistant
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US9431028B2 (en) 2010-01-25 2016-08-30 Newvaluexchange Ltd Apparatuses, methods and systems for a digital conversation management platform
US8977584B2 (en) 2010-01-25 2015-03-10 Newvaluexchange Global Ai Llp Apparatuses, methods and systems for a digital conversation management platform
US9424861B2 (en) 2010-01-25 2016-08-23 Newvaluexchange Ltd Apparatuses, methods and systems for a digital conversation management platform
US9424862B2 (en) 2010-01-25 2016-08-23 Newvaluexchange Ltd Apparatuses, methods and systems for a digital conversation management platform
US9190062B2 (en) 2010-02-25 2015-11-17 Apple Inc. User profiling for voice input processing
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US10049675B2 (en) 2010-02-25 2018-08-14 Apple Inc. User profiling for voice input processing
US8713021B2 (en) 2010-07-07 2014-04-29 Apple Inc. Unsupervised document clustering using latent semantic density analysis
US8719006B2 (en) 2010-08-27 2014-05-06 Apple Inc. Combined statistical and rule-based part-of-speech tagging for text-to-speech synthesis
US9075783B2 (en) 2010-09-27 2015-07-07 Apple Inc. Electronic device with text error correction based on voice recognition data
US8719014B2 (en) 2010-09-27 2014-05-06 Apple Inc. Electronic device with text error correction based on voice recognition data
US10515147B2 (en) 2010-12-22 2019-12-24 Apple Inc. Using statistical language models for contextual lookup
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US8781836B2 (en) 2011-02-22 2014-07-15 Apple Inc. Hearing assistance system for providing consistent human speech
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10102359B2 (en) 2011-03-21 2018-10-16 Apple Inc. Device access using voice authentication
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10255566B2 (en) 2011-06-03 2019-04-09 Apple Inc. Generating and processing task items that represent tasks to perform
US11120372B2 (en) 2011-06-03 2021-09-14 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10672399B2 (en) 2011-06-03 2020-06-02 Apple Inc. Switching between text data and audio data based on a mapping
US8812294B2 (en) 2011-06-21 2014-08-19 Apple Inc. Translating phrases from one language into another using an order-based set of declarative rules
US8706472B2 (en) 2011-08-11 2014-04-22 Apple Inc. Method for disambiguating multiple readings in language conversion
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US8762156B2 (en) 2011-09-28 2014-06-24 Apple Inc. Speech recognition repair using contextual information
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US8775442B2 (en) 2012-05-15 2014-07-08 Apple Inc. Semantic search using a single-source semantic model
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US10079014B2 (en) 2012-06-08 2018-09-18 Apple Inc. Name recognition system
US10019994B2 (en) 2012-06-08 2018-07-10 Apple Inc. Systems and methods for recognizing textual identifiers within a plurality of words
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US8935167B2 (en) 2012-09-25 2015-01-13 Apple Inc. Exemplar-based latent perceptual modeling for automatic speech recognition
US10978090B2 (en) 2013-02-07 2021-04-13 Apple Inc. Voice trigger for a digital assistant
US10199051B2 (en) 2013-02-07 2019-02-05 Apple Inc. Voice trigger for a digital assistant
US10572476B2 (en) 2013-03-14 2020-02-25 Apple Inc. Refining a search based on schedule items
US9977779B2 (en) 2013-03-14 2018-05-22 Apple Inc. Automatic supplementation of word correction dictionaries
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US9733821B2 (en) 2013-03-14 2017-08-15 Apple Inc. Voice control to diagnose inadvertent activation of accessibility features
US10642574B2 (en) 2013-03-14 2020-05-05 Apple Inc. Device, method, and graphical user interface for outputting captions
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US11388291B2 (en) 2013-03-14 2022-07-12 Apple Inc. System and method for processing voicemail
US10748529B1 (en) 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US10078487B2 (en) 2013-03-15 2018-09-18 Apple Inc. Context-sensitive handling of interruptions
US11151899B2 (en) 2013-03-15 2021-10-19 Apple Inc. User training by intelligent digital assistant
US9966060B2 (en) 2013-06-07 2018-05-08 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US10657961B2 (en) 2013-06-08 2020-05-19 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10185542B2 (en) 2013-06-09 2019-01-22 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
US20180012601A1 (en) * 2013-11-18 2018-01-11 Amazon Technologies, Inc. Dialog management with multiple applications
US11688402B2 (en) 2013-11-18 2023-06-27 Amazon Technologies, Inc. Dialog management with multiple modalities
US10706854B2 (en) * 2013-11-18 2020-07-07 Amazon Technologies, Inc. Dialog management with multiple applications
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US11257504B2 (en) 2014-05-30 2022-02-22 Apple Inc. Intelligent assistant for home automation
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9916831B2 (en) 2014-05-30 2018-03-13 Yandex Europe Ag System and method for handling a spoken user request
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US10497365B2 (en) 2014-05-30 2019-12-03 Apple Inc. Multi-command single utterance input method
US11133008B2 (en) 2014-05-30 2021-09-28 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US10083690B2 (en) 2014-05-30 2018-09-25 Apple Inc. Better resolution when referencing to concepts
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US10169329B2 (en) 2014-05-30 2019-01-01 Apple Inc. Exemplar-based natural language processing
RU2654789C2 (en) * 2014-05-30 2018-05-22 Общество С Ограниченной Ответственностью "Яндекс" Method (options) and electronic device (options) for processing the user verbal request
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10904611B2 (en) 2014-06-30 2021-01-26 Apple Inc. Intelligent automated assistant for TV user interactions
US9668024B2 (en) 2014-06-30 2017-05-30 Apple Inc. Intelligent automated assistant for TV user interactions
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US10431204B2 (en) 2014-09-11 2019-10-01 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US9986419B2 (en) 2014-09-30 2018-05-29 Apple Inc. Social reminders
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10210003B2 (en) * 2014-09-30 2019-02-19 Nuance Communications, Inc. Methods and apparatus for module arbitration
US20170308389A1 (en) * 2014-09-30 2017-10-26 Nuance Communications, Inc. Methods And Apparatus For Module Arbitration
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US11556230B2 (en) 2014-12-02 2023-01-17 Apple Inc. Data detection
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US11087759B2 (en) 2015-03-08 2021-08-10 Apple Inc. Virtual assistant activation
US10311871B2 (en) 2015-03-08 2019-06-04 Apple Inc. Competing devices responding to voice triggers
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10356243B2 (en) 2015-06-05 2019-07-16 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US11500672B2 (en) 2015-09-08 2022-11-15 Apple Inc. Distributed personal assistant
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US11526368B2 (en) 2015-11-06 2022-12-13 Apple Inc. Intelligent automated assistant in a messaging environment
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US11069347B2 (en) 2016-06-08 2021-07-20 Apple Inc. Intelligent automated assistant for media exploration
US10354011B2 (en) 2016-06-09 2019-07-16 Apple Inc. Intelligent automated assistant in a home environment
US11037565B2 (en) 2016-06-10 2021-06-15 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10733993B2 (en) 2016-06-10 2020-08-04 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US11152002B2 (en) 2016-06-11 2021-10-19 Apple Inc. Application integration with a digital assistant
US10521466B2 (en) 2016-06-11 2019-12-31 Apple Inc. Data driven natural language event detection and classification
US10089072B2 (en) 2016-06-11 2018-10-02 Apple Inc. Intelligent device arbitration and control
US10297253B2 (en) 2016-06-11 2019-05-21 Apple Inc. Application integration with a digital assistant
US10269345B2 (en) 2016-06-11 2019-04-23 Apple Inc. Intelligent task discovery
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10553215B2 (en) 2016-09-23 2020-02-04 Apple Inc. Intelligent automated assistant
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US10755703B2 (en) 2017-05-11 2020-08-25 Apple Inc. Offline personal assistant
US11405466B2 (en) 2017-05-12 2022-08-02 Apple Inc. Synchronization and task delegation of a digital assistant
US10791176B2 (en) 2017-05-12 2020-09-29 Apple Inc. Synchronization and task delegation of a digital assistant
US10410637B2 (en) 2017-05-12 2019-09-10 Apple Inc. User-specific acoustic models
US10810274B2 (en) 2017-05-15 2020-10-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US10482874B2 (en) 2017-05-15 2019-11-19 Apple Inc. Hierarchical belief states for digital assistants
US11217255B2 (en) 2017-05-16 2022-01-04 Apple Inc. Far-field extension for digital assistant services

Also Published As

Publication number Publication date
US6944594B2 (en) 2005-09-13
WO2002097795A1 (en) 2002-12-05
US20020184023A1 (en) 2002-12-05

Similar Documents

Publication Publication Date Title
US6944594B2 (en) Multi-context conversational environment system and method
US6496799B1 (en) End-of-utterance determination for voice processing
US6601029B1 (en) Voice processing apparatus
US6438520B1 (en) Apparatus, method and system for cross-speaker speech recognition for telecommunication applications
US7552054B1 (en) Providing menu and other services for an information processing system using a telephone or other audio interface
JP3561076B2 (en) Automatic call recognition method for arbitrarily spoken words
US6246986B1 (en) User barge-in enablement in large vocabulary speech recognition systems
Kamm User interfaces for voice applications.
EP0780829B1 (en) Method for automatic speech recognition in telephony
US7783475B2 (en) Menu-based, speech actuated system with speak-ahead capability
US6282511B1 (en) Voiced interface with hyperlinked information
US7308408B1 (en) Providing services for an information processing system using an audio interface
US7260537B2 (en) Disambiguating results within a speech based IVR session
US6687673B2 (en) Speech recognition system
US5797124A (en) Voice-controlled voice mail having random-order message retrieval based on played spoken identifier list
US6882973B1 (en) Speech recognition system with barge-in capability
EP1976255B1 (en) Call center with distributed speech recognition
JP2000078296A (en) Direction method in interactive sound response system and the interactive sound response system
US5930336A (en) Voice dialing server for branch exchange telephone systems
US20030191639A1 (en) Dynamic and adaptive selection of vocabulary and acoustic models based on a call context for speech recognition
US20040215451A1 (en) Telephone call handling center where operators utilize synthesized voices generated or modified to exhibit or omit prescribed speech characteristics
JPH10513033A (en) Automatic vocabulary creation for voice dialing based on telecommunications networks
JP2003520983A (en) Improved text-to-speech conversion
GB2348035A (en) Speech recognition system
US8145495B2 (en) Integrated voice navigation system and method

Legal Events

Date Code Title Description
AS Assignment

Owner name: AT&T INTELLECTUAL PROPERTY I, L.P., NEVADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AT&T DELAWARE INTELLECTUAL PROPERTY, INC.;REEL/FRAME:022266/0765

Effective date: 20090213

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: BELLSOUTH INTELLECTUAL PROPERTY CORPORATION, DELAW

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BUSAYAPONGCHAI, SENIS;CHINTRAKULCHAI, PICHET;REEL/FRAME:034429/0037

Effective date: 20010730

Owner name: AT&T BLS INTELLECTUAL PROPERTY, INC., DELAWARE

Free format text: CHANGE OF NAME;ASSIGNOR:AT&T INTELLECTUAL PROPERTY, INC.;REEL/FRAME:034540/0202

Effective date: 20070727

Owner name: AT&T INTELLECTUAL PROPERTY, INC., TEXAS

Free format text: CHANGE OF NAME;ASSIGNOR:BELLSOUTH INTELLECTUAL PROPERTY CORPORATION;REEL/FRAME:034540/0192

Effective date: 20070427

Owner name: AT&T DELAWARE INTELLECTUAL PROPERTY, INC., DELAWAR

Free format text: CHANGE OF NAME;ASSIGNOR:AT&T BLS INTELLECTUAL PROPERTY, INC.;REEL/FRAME:034540/0207

Effective date: 20071101

AS Assignment

Owner name: AT&T ALEX HOLDINGS, LLC, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AT&T INTELLECTUAL PROPERTY I, L.P.;REEL/FRAME:034467/0853

Effective date: 20141210

AS Assignment

Owner name: NUANCE COMMUNICATIONS, INC., MASSACHUSETTS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AT&T ALEX HOLDINGS, LLC;REEL/FRAME:041495/0903

Effective date: 20161214

AS Assignment

Owner name: AT&T INTELLECTUAL PROPERTY I, L.P., GEORGIA

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE/ASSIGNOR PREVIOUSLY RECORDED ON REEL 034467 FRAME 0853. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:AT&T DELAWARE INTELLECTUAL PROPERTY, INC.;REEL/FRAME:042961/0928

Effective date: 20090213