US20080221892A1 - Systems and methods for an autonomous avatar driver - Google Patents

Systems and methods for an autonomous avatar driver Download PDF

Info

Publication number
US20080221892A1
US20080221892A1 US11/960,507 US96050707A US2008221892A1 US 20080221892 A1 US20080221892 A1 US 20080221892A1 US 96050707 A US96050707 A US 96050707A US 2008221892 A1 US2008221892 A1 US 2008221892A1
Authority
US
United States
Prior art keywords
dialog
language
semantic
avatar
generate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/960,507
Inventor
Paco Xander Nathan
Cathi Joann Cox
Florian Thomas Leibert
Mark Stephen Meadows
Jan Susan Mallis
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Botanic Technologies Inc
OGG LLC
Original Assignee
HEADCASE HUMANUFACTURING Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by HEADCASE HUMANUFACTURING Inc filed Critical HEADCASE HUMANUFACTURING Inc
Priority to US11/960,507 priority Critical patent/US20080221892A1/en
Assigned to HEADCASE HUMANUFACTURING, INC. reassignment HEADCASE HUMANUFACTURING, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MEADOWS, MARK STEPHEN, NATHAN, PACO XANDER, COX, CATHI JOANN, LEIBERT, FLORIAN THOMAS, MALLIS, JAN SUSAN
Publication of US20080221892A1 publication Critical patent/US20080221892A1/en
Assigned to OGG, LLC reassignment OGG, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HEADCASE HUMANUFACTURING, INC.
Assigned to MEADOWS, MARK STEPHEN reassignment MEADOWS, MARK STEPHEN ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OGG, LLC
Priority to US14/536,626 priority patent/US10133733B2/en
Assigned to BOTANIC TECHNOLOGIES, INC. reassignment BOTANIC TECHNOLOGIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MEADOWS, MARK STEPHEN
Assigned to BOTANIC TECHNOLOGIES, INC. reassignment BOTANIC TECHNOLOGIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HEADCASE HUMANUFACTURING, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • G06F40/35Discourse or dialogue representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/263Language identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation

Definitions

  • the present invention relates to a system and method for autonomous avatar driver, and more particularly an autonomous avatar driver capable of natural language processing for improved data querying and enhanced knowledge representation for avatars, virtual healthcare advisors, virtual personal assistants and non player characters within video games, Massively Multiplayer Online Games (MMOGs), virtual worlds, online social networks, and virtual classrooms.
  • the autonomous avatar driver may also be enabled to mine dialog for data that may then be utilized to generate recommendations to a user.
  • the number of active subscribers to MMOGs is at least 10 million people. Each person pays $15 and up a month to play these games, and maybe and additional 20 million people login occasionally. Estimates are that players spent about $1 billion in real money in 2005 on virtual goods and services for MMOGs combined. Moreover, at least 1.5 million people subscribe to virtual worlds. In January, 2006, inside one such virtual social world, people spent nearly $5 million in some 4.2 million transactions buying or selling clothes, buildings, and the like. Moreover, participants in web communities number in the multiple tens of millions. Additionally, traditional video games have sold over $30 billion in goods since 2000.
  • knowledge representation is fatally limited by resource and financial shortcomings.
  • Data required for knowledge representation may be incorporated into databases in particular formats to be usable for knowledge representation.
  • Such a system is cumbersome, requiring large numbers of man-hours, often with costly specialists, and huge amounts of storage for the data.
  • the data set created for the knowledge representation will be incomplete, leading to palpable gaps in the “knowledge” of the character.
  • a natural language process that is able to examine external data sources, and glean relevant data may provide far more cost efficient, complete and adaptable data sources for knowledge representation is currently lacking.
  • knowledge representation may become essential to the success of the product.
  • personal avatars may likewise benefit from knowledge representation as knowledgeable personal assistants, and more fully developed player characters.
  • the present invention allows for this level of expanded knowledge representation.
  • a natural language interface may allow substantially improved searching for relevant frequently asked questions and general data queries. Such improvements have considerable implications for education, healthcare and corporate relations. Current data queries examine matches of words rather than matches of meaning. Natural language processing may improve these searching methods by incorporating semantics.
  • Such advanced contextual search techniques may be extended to avatar dialog to generate highly targeted data mining.
  • Such data mining may be useful for advertisement, user recommendations and statistical data farming.
  • systems and methods for an autonomous avatar driver are provided. Such systems and methods are useful for providing a highly believable virtual avatar, or character, with interactive ability and knowledge representation and recommendation ability.
  • the autonomous avatar driver is useful in association with language sources.
  • a sourcer may receive dialog from the language source. It may also, in some embodiments, receive external data from data sources.
  • a segmentor may segment the dialog.
  • the segmentor may include a particle representor, a character converter and a splitter for splitting dialog.
  • a parser may then parse the segmented dialog.
  • the parser may include a linker for linking grammar, a grammatical mood analyzer, a tagger for tagging the dialog and a variant pruner.
  • the semantic engine may analyze the parsed dialog for semantics.
  • the semantics engine may include a token framer, a lemmatizer configured to draw from semantic lexicons, a semantic network generator and a co-reference resolver for resolving co-reference ambiguity.
  • An analytics engine may then analyze the semantically analyzed dialog for analytics.
  • the analytics engine may include a common word filter, a token frame statistics engine, and an N-gram analyzer configured to analyze N-grams for word popularity.
  • the token frame statistics engine may count lemmatized words, and perform nodal analysis.
  • a pragmatics analyzer may analyze the analytically analyzed dialog for pragmatics.
  • the pragmatics analyzer may include a slang resolver, a knowledge template resolver configured to generate knowledge templates, a proper noun grouper and an affect estimator configured to estimate emotion of the dialog.
  • a recommender may then generate recommendations from the pragmatically analyzed dialog.
  • the recommender may include a cloud aggregator for generating aggregate tag clouds, a hood cluster or for clustering the end users into neighborhoods based upon dialog parsed from them, a social network recommender for recommending social networking to individuals and businesses, and a contextual advertising generator.
  • a response generator may generate responses for the autonomous avatar using the pragmatically analyzed dialog.
  • the response generator may also incorporate the generated recommendations from the recommender.
  • FIG. 1A shows a schematic block diagram illustrating an autonomous avatar driver with a natural language processor system with knowledge representation in accordance with some embodiments of the present invention
  • FIG. 1B shows an exemplary functional block diagram of the natural language processor in accordance with some embodiments of the present invention
  • FIG. 2 shows an exemplary functional block diagram of the natural language analyzer of FIG. 1 ;
  • FIG. 3 shows a flow chart illustrating the natural language processing for the natural language processor system of FIG. 1 ;
  • FIG. 4 shows a flow chart illustrating the process for processing language for knowledge representation and semantics for the natural language processor system of FIG. 1 ;
  • FIG. 5 shows a flow chart illustrating the process for preparing language for the natural language processor system of FIG. 1 ;
  • FIG. 6 shows a flow chart illustrating the process for chunking text for the natural language processor system of FIG. 1 ;
  • FIG. 7 shows a flow chart illustrating the process for setting grammatical mood for the natural language processor system of FIG. 1 ;
  • FIG. 8 shows a flow chart illustrating the process for pattern recognition for the natural language processor system of FIG. 1 ;
  • FIG. 9 shows a flow chart illustrating the process for post operations for the natural language processor system of FIG. 1 ;
  • FIG. 10 shows a flow chart illustrating the process for preparing processed text for retrieval for the natural language processor system of FIG. 1 ;
  • FIG. 11 shows a flow chart illustrating the process for an embodiment of data preparation for semantic searches utilizing the natural language processor system of FIG. 1 ;
  • FIG. 12 shows a flow chart illustrating the process for an embodiment of semantic searches for frequently-asked-questions utilizing the natural language processor system of FIG. 1 ;
  • FIG. 13A shows a flow chart illustrating the process for one embodiment of determining semantic similarities for semantic searches utilizing the natural language processor system of FIG. 1 ;
  • FIG. 13B shows a flow chart illustrating the process for another embodiment of determining semantic similarities for semantic searches utilizing the natural language processor system of FIG. 1 ;
  • FIG. 14 shows a flow chart illustrating the process for an embodiment of a general semantic search utilizing the natural language processor system of FIG. 1 ;
  • FIG. 15A shows a flow chart illustrating the process for one embodiment of determining semantic similarities for general semantic searches utilizing the natural language processor system of FIG. 1 ;
  • FIG. 15B shows a flow chart illustrating the process for another embodiment of determining semantic similarities for general semantic searches utilizing the natural language processor system of FIG. 1 ;
  • FIG. 16 shows a flow chart illustrating the process for generating tag clouds in accordance with an embodiment of the present invention
  • FIG. 17 shows a schematic block diagram illustrating an autonomous avatar driver system with recommendation ability in accordance with some embodiments of the present invention.
  • FIG. 18 shows a schematic block diagram illustrating a sourcer in accordance with the autonomous avatar driver system of FIG. 17 ;
  • FIG. 19 shows a schematic block diagram illustrating a segmenter in accordance with the autonomous avatar driver system of FIG. 17 ;
  • FIG. 20 shows a schematic block diagram illustrating a parser in accordance with the autonomous avatar driver system of FIG. 17 ;
  • FIG. 21 shows a schematic block diagram illustrating a semantic engine in accordance with the autonomous avatar driver system of FIG. 17 ;
  • FIG. 22 shows a schematic block diagram illustrating an analytics engine in accordance with the autonomous avatar driver system of FIG. 17 ;
  • FIG. 23 shows a schematic block diagram illustrating a pragmatics analyzer in accordance with the autonomous avatar driver system of FIG. 17 ;
  • FIG. 24 shows a schematic block diagram illustrating a recommender in accordance with the autonomous avatar driver system of FIG. 17 ;
  • FIG. 25 shows a schematic block diagram illustrating a response generator in accordance with the autonomous avatar driver system of FIG. 17 ;
  • FIG. 26A shows a schematic block diagram of the virtual universe for the autonomous avatar driver system of FIG. 17 ;
  • FIG. 26B shows a logical block diagram of virtual environments for the autonomous avatar driver system of FIG. 17 ;
  • FIG. 26C shows a schematic block diagram of a virtual environment for the autonomous avatar driver system of FIG. 17 ;
  • FIG. 26D shows a schematic block diagram of an autonomous avatar for the autonomous avatar driver system of FIG. 17 ;
  • FIG. 27 shows a flow chart illustrating the process for driving the autonomous avatar in accordance with some embodiments of the present invention
  • FIG. 28 shows a flow chart illustrating the process for dialog segmentation in accordance with some embodiments of the present invention.
  • FIG. 29 shows a flow chart illustrating the process for dialog parsing in accordance with some embodiments of the present invention.
  • FIG. 30 shows a flow chart illustrating the process for semantic analysis in accordance with some embodiments of the present invention.
  • FIG. 31 shows a flow chart illustrating the process for analytic analysis in accordance with some embodiments of the present invention.
  • FIG. 32 shows a flow chart illustrating the process for pragmatic analysis in accordance with some embodiments of the present invention.
  • FIG. 33 shows a flow chart illustrating the process for recommendation generation in accordance with some embodiments of the present invention.
  • FIG. 34 shows a flow chart illustrating the process for avatar dialog generation in accordance with some embodiments of the present invention.
  • the present invention relates to systems and methods for autonomous avatar driver that provides for enhanced functionality, utility and believability for avatars, and recommendation ability.
  • Such autonomous avatar drivers are useful for generating believable avatars in conjunction with Massively Multiplayer Online Games (MMOGs), virtual social worlds, online web communities, corporate web sites, health care and educational programs and websites.
  • MMOGs Massively Multiplayer Online Games
  • Believability of the avatars is due to their ability to respond to language, also referred to as dialogue, also known in the art as “natural language”, in a spontaneous manner, and their ability to provide an extensive knowledge base, referred to as knowledge representation.
  • natural language processors allow for semantic searches, searches that query not only the words provided but the meaning, thereby making the search more efficient, intuitive and useful.
  • an avatar is a virtual representation of an individual within a virtual environment.
  • Avatars often include physical characteristics, statistical attributes, inventories, social relations, emotional representations, and weblogs (blogs) or other recorded historical data.
  • Avatars may be human in appearance, but are not limited to any appearance constraints.
  • Avatars may be personifications of a real world individual, such as a Player Character (PC) within a MMOG, or may be an artificial personality, such as a Non-Player Character (NPC).
  • PC Player Character
  • NPC Non-Player Character
  • Additional artificial personality type avatars include personal assistants, guides, educators, answering servers and information providers. Additionally, some avatars may have the ability to be automated some of the time, and controlled by a human at other times. Such Quasi-Player Characters (QPCs) may perform mundane tasks automatically, but more expensive human agents take over in cases of complex problems.
  • QPCs Quasi-Player Characters
  • the avatar driven by the autonomous avatar driver is generically defined.
  • the avatar may be a character, non-player character, quasi-player character, agent, personal assistant, personality, guide, representation, educator or any additional virtual entity within virtual environments.
  • Avatars may be as complex as a 3D rendered graphical embodiment that includes detailed facial and body expressions, or may be as simple as a faceless, non-graphical widget, capable of limited, or no function beyond the natural language processor. In a society of ever increasing reliance and blending between real life and our virtual lives, the ability to have believable and useful avatars is highly desirable and advantageous.
  • FIG. 1A shows a schematic block diagram illustrating an Autonomous Avatar Driver System 100 with knowledge representation in accordance with some embodiments of the present invention.
  • the Autonomous Avatar Driver 150 Central to the Autonomous Avatar Driver System 100 is the Autonomous Avatar Driver 150 .
  • the Autonomous Avatar Driver 150 and Natural Language Processor 190 are seen coupled to a Local Area Network (LAN) 106 in a network operations center (NOC).
  • NOC network operations center
  • the Natural Language Processor 190 and/or Autonomous Avatar Driver 150 may be located at Language Sources 103 a to 103 m .
  • the Autonomous Avatar Driver 150 and/or Natural Language Processor 190 may be located at hosting systems. Host systems may include corporate servers, game systems, MMOG's, websites or other virtual environments.
  • the Language Sources 103 a , 103 b to 103 m may be connected to the LAN 106 through a Wide Area Network (WAN) 101 .
  • WAN Wide Area Network
  • the most common WAN 101 is the internet; however any WAN, such as a closed gamming network, is intended to be considered as within the scope of the WAN 101 .
  • Language Sources 103 p , 103 q to 103 z may couple directly to the LAN 106 .
  • Language Sources 103 a to 103 z provide language to the Natural Language Processor 190 for processing.
  • Language may be in the form of text, or in some embodiments may include audio files and graphical representations, including facial features and body language.
  • the Natural Language Processor 190 includes a Natural Language Analyzer 130 , Data Miner 120 , Databases 110 and Natural Language Generator 140 .
  • a Firewall 104 separates the WAN 101 and Language Sources 103 p , 103 q to 103 z from the LAN 106 .
  • the Firewall 104 is coupled to a Server 105 , which acts as an intermediary between the Firewall 104 and the LAN 106 .
  • Data Sources 102 a , 102 b to 102 x may be coupled to the WAN 101 for access by the Autonomous Avatar Driver 150 and/or Natural Language Processor 190 .
  • Data Sources 102 a to 102 x provide external data for use by the Autonomous Avatar Driver 150 and/or Natural Language Processor 190 for the knowledge representation.
  • Databases 110 may be coupled to the LAN 106 with for storing frequently utilized data, and for dialogue storage. Together, Data Sources 102 a to 102 x and the Databases 110 provide the statistical information and data required for accurate language processing and knowledge representation.
  • the Databases 110 may include, but are not limited to four child databases.
  • These databases include a dictionary, wordnet, part-of-speech tags (or speech tags), and conversation history.
  • These Databases 110 provide the resources necessary for basic natural language processing with a persistent conversation memory. Additional databases may be included as is desired, since it is the intention of the invention to be amorphic and adaptable as technology and resources become available.
  • the Autonomous Avatar Driver 150 as disclosed may be coupled to the Natural Language Analyzer 130 , Data Miner 120 and Natural Language Generator 140 . In some embodiments, the Autonomous Avatar Driver 150 may utilize these components to drive the avatar. In some alternate embodiments, the Autonomous Avatar Driver 150 may include alternate components that are capable of language analysis and response.
  • the Natural Language Analyzer 130 may be coupled to a Data Miner 120 and Natural Language Generator 140 .
  • the Natural Language Analyzer 130 generates a wealth of data regarding conversations including the meaning, moods and responses to the language, or dialogue. Such data may be extremely valuable for the generation of highly targeted advertising, statistical information, marketing data, sociological research data, political barometer, and many other areas of interest.
  • the Data Miner 120 allows for the searching and congregation of said content-dense data for these purposes.
  • the Natural Language Generator 140 is capable of receiving semantics, in the form of lexical chains, grammatical moods and represented knowledge, to generate an appropriate response. In some embodiments, the Natural Language Generator 140 returns text. In alternate embodiments the Natural Language Generator 140 may return a response in the native form of the Language Source 103 a to 103 z . Thus, in some embodiments voice synthesizer capabilities may be included in the Natural Language Generator 140 . In some embodiments, it is advantageous to include a system within the Natural Language Generator 140 that is capable of identifying taboo or undesired language and prevent such language from being part of responses. This feature is very important when dealing with children, or other individuals sensitive to particular conversations.
  • the Natural Language Generator 140 may, in some embodiments, be designed to incorporate personalized product placement for advertising purposes when relevant to the conversation. In fact, in some embodiments, the Natural Language Generator 140 may be designed to steer conversations toward product placement, political views, etc. In such a way, avatars incorporating such a Natural Language Processor 190 may become highly specialized and sophisticated advertising media.
  • the Natural Language Generator 140 may include an accompanying graphical generator.
  • Said graphical generator may provide graphical representations, movements, facial features, stance, gestures or other visual stimulus appropriate to the generated language.
  • Such a system may be desirable for very personalized interaction; however, such a system requires large computations and bandwidth to be practical for multiple avatars.
  • said system for graphical generation is best utilized when the Autonomous Avatar Driver System 100 is hosted on an individual's system in order to run a personal avatar, or on a corporate server for the corporate avatar.
  • the graphical generator may output a set of mood indicators that may then be utilized by the avatar host to generate movements that correspond to the given mood of the language. For example a series of predetermined mood categories may be assigned, and each mood category includes a numerical tag. The host system may have postures, facial features and physical movements corresponding to each of the mood categories. The graphical generator outputs the numerical mood tag, and the host system is then able to produce a range of graphics appropriate to the language. Such a system minimizes computations required by the graphical generator and minimizes required bandwidth. Moreover, in some embodiments, the graphical generator may provide a series of graphical indicators to more finely tune graphical representations.
  • the graphical generator may output a first variable relating to the general mood, and subsequent variables, each more finely tuning the graphical representations.
  • the advantage of such a system is that the complexity of host graphical representations may vary greatly in their level of detail and range of emotional states. As such, the host system may utilize outputs, and ignore subsequent tuning variables as they become ineffectual within said host system.
  • the advantage of such a system is the ability of a standard output by the graphical generator, yet having each host system receiving the level of graphical detail desired.
  • FIG. 1B shows an exemplary functional block diagram of the Natural Language Processor 190 as involved in the Autonomous Avatar Driver System 100 of FIG. 1 .
  • Exemplary Language Source 130 a may be seen providing language to the Natural Language Processor 190 .
  • the language is received by the Natural Language Analyzer 130 , which analyzes the language for semantics.
  • the Natural Language Analyzer 130 is coupled to the Natural Language Generator 140 , the Databases 110 and the Data Miner 120 .
  • the exemplary Data Source 102 a provides data to the Natural Language Analyzer 130 .
  • the Natural Language Analyzer 130 may utilize the Databases 110 for analyzing the language semantics.
  • the Natural Language Analyzer 130 may provide semantic information from the analyzed language to the Data Miner 120 and the Natural Language Generator 140 .
  • the Natural Language Generator 140 is coupled to the Natural Language Analyzer 130 , the Databases 110 and the Data Miner 120 . Moreover, the exemplary Data Source 102 a provides data to the Natural Language Generator 140 . The Natural Language Generator 140 may utilize the Databases 110 for generating return language. Additionally, the Natural Language Generator 140 may provide semantic information from the generated language to the Data Miner 120 . The generated language may be provided to the exemplary Language Source 130 a . In this way, the Natural Language Processor 190 is capable of semantically relevant conversation, with knowledge representation.
  • the Data Miner 120 aside from gaining semantic information from the Natural Language Generator 140 and Natural Language Analyzer 130 may query the Databases 110 in some embodiments.
  • the Data Miner 120 is capable of providing the mined data to Data Consumers 125 .
  • Data Consumers 125 include research groups, advertisers, educators, marketing analysts, political analysts, quality control analysts, and opinion based review groups, among many others.
  • Such Data Consumers 125 may provide an important source of revenue in some embodiments, as well as providing service data for the end users.
  • FIG. 2 shows a functional block diagram of the Natural Language Analyzer 130 of FIG. 1 .
  • a ‘lemma’ is a set of lexical forms with the same stem, the same part-of-speech and the same word-sense.
  • a ‘lexeme’ is a pairing of a particular orthographic form with a symbolic meaning representation.
  • a finite list of lexemes is referred to as a ‘lexicon’.
  • ‘types’ refers to the total number of distinct words in a corpus.
  • a Receiver 201 receives the Language from the Language Source 103 a to 103 z from over the network.
  • the Receiver 201 may also identify the format of the language.
  • the Receiver 201 may include a spelling checker; however, as it will be seen, improper grammar, slang and typos are anticipated and dealt with by the Natural Language Analyzer 130 .
  • the Receiver 201 is coupled to the Language Preparer 220 which includes the Language Converter 221 , Sentence Detector 222 , Part-of-Speech Tagger (or Speech Tagger) 223 and Speech Parser (or Parser) 224 .
  • the Receiver 201 couples directly with the Language Converter 221 and the Sentence Detector 222 .
  • the Receiver 201 may provide the language directly to the Sentence Detector 222 .
  • the native language for the Natural Language Analyzer 130 includes text files.
  • the dialog is first analyzed by the Language Converter 221 .
  • the Language Converter 221 may receive a plurality of language formats and convert them to a text format. In said Language Converter 221 , voice recognition software may be required.
  • the Language Converter 221 may include image recognition in order to interpret body language, sign language and facial features for conversion into the native language for the Natural Language Analyzer 130 . After nonnative language is analyzed by the Language Converter 221 the text equivalent of the language is processed by the Sentence Detector 222 .
  • the instant invention is designed to be able to analyze, mine and generate in a variety of languages.
  • the base architecture of the Autonomous Avatar Driver System 100 is nearly identical regardless of whether the Language Sources 103 a to 103 z are in English, German or Japanese, to name a few applicable languages.
  • the instant invention utilizes statistical learning, dictionaries and grammatical rules; as such, as long as the appropriate training corpuses are available the system may be utilized across a variety of languages.
  • a translation matrix may be utilized in order to provide semantic consistency, and when it is desirous for the input language format to differ from the output language format.
  • Such a translation matrix may exist on the front end, translating the incoming language.
  • said translation matrix may exist after the Natural Language Analyzer 130 in order to translate the semantic meaning of the received language.
  • said translation matrix may exist before the Natural Language Generator 140 , thereby enabling all analysis to be performed within the native language format, and only language generation is performed in the output language format.
  • said translation matrix may exist after the Natural Language Generator 140 , thereby enabling all natural language processing to be performed within the native language format, and only language output is in the output language format.
  • a combination of translation matrices may be utilized in some embodiments.
  • the Sentence Detector 222 is coupled with the Language Converter 221 and the Part-of-Speech Tagger (or Speech Tagger) 223 .
  • the Sentence Detector 222 utilizes algorithms to determine separate sentences.
  • the sentences are made up of elements, which include words and symbols.
  • the Speech Tagger 223 first cleans up the raw text by tokenizing it. Tokenization includes replacing certain elements with tokens. The tokens delineate meanings of the elements that they replace. Additionally, tokenization ensures that special characters are separated from the words in the sentence so that the elements may be parsed easier.
  • the Speech Tagger 223 then annotates each element with a tag that classifies the element with pre-defined categories.
  • the Speech Tagger 223 is usually trained with a training set of data since it relies upon statistics in order to determine the meaning of ambiguous words.
  • the Speech Tagger 223 may output the tagged sentence via the Outputter 203 , or may continue through the Speech Parser 224 , which is coupled to the Speech Tagger 223 .
  • Speech Parser 224 chunks, or parses, the sentences into discrete, non-overlapping grammatical chunks. Each chunk is a partial structure of the sentence. Chunking is important for downstream semantic analysis and pattern recognition. After Chunking is performed, the chunks are annotated for property based searching and improved information extraction. Examples of chunking rules utilized by the Speech Parser 224 are illustrated below in Table 2:
  • the Speech Parser 224 may output the chunked sentence via the Outputter 203 or may continue through the Semantic Analyzer 230 , which is coupled to the Speech Parser 224 .
  • the Semantic Analyzer 230 includes a Grammar Analyzer 231 , a Referencer 232 , an Affect Analyzer 233 and a Pattern Recognizer 234 .
  • the Grammar Analyzer 231 is coupled to the Speech Parser 224 , the Referencer 232 and the Outputter 203 .
  • Grammar Analyzer 231 determines the grammatical mood of the sentence. Examples of grammatical moods assigned by the Grammar Analyzer 231 are illustrated below in Table 3:
  • additional definitions of grammatical moods may be utilized by the Grammar Analyzer 231 ; however in some other embodiments the grammatical moods illustrated in Table 3 are sufficient, and any dialogue that does not fit into one of the previous definitions may be assumed to be a declarative grammatical mood. Such an assumption generally proves computationally effective, as well as accurate enough for most circumstances.
  • the Grammar Analyzer 231 may utilize a statistical engine, which may incorporate training, in order to statistically determine the grammatical mood, as apposed to a rule based engine as described above.
  • a hybrid rule and statistical engine may be utilized, wherein the rules are utilized for simple sentence structures, but learned statistical methods may be utilized for more complicated or less determinable sentences.
  • Such a system may be of particular use when slang is highly prevalent within the language. The advantages of such a system include the incorporation of learning, or statistical tuning, in order to improve accuracy.
  • the Grammar Analyzer 231 is also coupled to the Outputter 203 , such as to output the grammatical mood of the sentence. Grammatical moods may result in prioritization of discourse; for example imperative structured sentences may require an action or trigger an event.
  • the Referencer 232 may reference dictionaries and thesauruses in order to reference the elements of the sentence for semantics. Additionally, in some embodiments, the Referencer 232 may cross reference graphical cues, such as gestures, with language pronouns in order to delineate the meaning of ambiguous pronouns. The Referencer 232 is then coupled to the Affect Analyzer 233 .
  • the Affect Analyzer 233 utilizes the results of the Referencer 232 and the grammatical mood from the Grammar Analyzer 231 in order to infer the affect of the sentence. Graphical recognition may be of particular use by the Affect Analyzer 233 to infer the language affect. Additionally, in some embodiments, tactic responses may be utilized to analyze affect. Tactic responses, as is known to those skilled in the art, includes the usage of information, such as the rate of typing, change in voice pitch, thermal readings of skin, electrical conductivity of skin, etc. in order to help infer the users' emotional state for the purposes of affect analysis. Such tactic response devices may be as simple as analyzing the rhythm of typing, or may include external peripheral devices worn by the users.
  • the Affect Analyzer 233 may provide cues as to the appropriate disposition of return language to the Natural Language Generator 140 . Results from the Affect Analyzer 233 may proceed directly to the Knowledge Representor 202 or may proceed through the Pattern Recognizer 234 .
  • Pattern Recognizer 234 utilizes statistics, through a statistical engine, to determine meaning of the language. As such, the Pattern Recognizer 234 may be trained. In some embodiments it may be advantageous to train the Pattern Recognizer 234 within the environment that it will function, in order to incorporate the slang of that environment. Within the gamming community entire dialects of speech have developed, often differing between game to game, and even separate factions within single games. It would be advantageous to have the Natural Language Analyzer 130 enabled to process slang, and the Natural Language Generator 140 configured to use said slang, in order to enhance the believability of the avatar. Additionally, it is important to note that these dialects and slang language adapt rapidly, therefore a constant training of the Pattern Recognizer 234 may be beneficial.
  • the Pattern Recognizer 234 may generate a lexical chain in order to convey the meaning of the language dialogue.
  • Lexical chains well known to those skilled in the art, are strings of words designed to convey meanings of bodies of text, and are designed to eliminate incongruities and ambiguities.
  • the lexical chains generated by the Pattern Recognizer 234 may be processed through the Knowledge Representor 202 .
  • the Knowledge Representor 202 utilizes knowledge templates in order to query Data Sources 102 a to 102 x and Databases 110 in order to provide a more enriching and believable conversation with the avatar.
  • the Databases 110 may include previous conversation logs, providing a level of personal ‘memory’ for the avatar. In this way, previous conversations and happenings may be ‘recalled’ by the avatar.
  • external Data Sources 102 a to 102 x may provide a vast wealth of information about virtually any topic, upon which the avatar may draw from in order to effectuate a believable dialogue.
  • Data Sources 102 a to 102 x may include reference sources such a Wikipedia® and IMDb®.
  • knowledge templates may be uniquely designed to access a particular Data Sources 102 a to 102 x .
  • any topic contained within the Data Sources 102 a to 102 x may be analyzed by the knowledge template to provide meaningful information for the Knowledge Representor 202 of the Natural Language Analyzer 130 .
  • the results of the Knowledge Representor 202 may be sent to the Outputter 203 for downstream applications.
  • FIG. 3 shows a flow chart illustrating the natural language processing shown generally at 300 .
  • step 310 the language received from the Language Source 103 a to 103 z is analyzed for knowledge representation and semantics by the Natural Language Analyzer 130 .
  • step 315 a decision is made whether to perform data mining on the analyzed language. If data mining is desired, then the semantics of the analyzed language are mined for the pertinent data, at step 320 . Such data may then be stored for future uses or supplied to the Data Consumers 125 .
  • the process then proceeds to step 325 , where a decision is made whether to store the results of the natural language analysis.
  • step 315 data mining is not desired, the process proceeds directly to step 325 , where a decision is made whether to store the results of the natural language analysis. If it is decided to store the natural language analysis at step 325 , then the semantics may be stored, or archived, within the Databases 110 at step 330 . Stored analyzed language may be subsequently data mined. Also, in some embodiments, the stored analyzed language may be retrieved in future interactions, providing an avatar with a ‘memory’ of previous conversations, and thus resulting in enhanced believability of the avatar. Also, the stored analyzed language may be subsequently cross referenced in order to provide semantics for future ambiguities.
  • ambiguous pronouns in future language may rely upon the previous language semantics in order to resolve said ambiguity.
  • these stored semantics may be annotated with the time, location, identity of Language Source 103 a to 103 z and any additional relevant information.
  • the entire corpus of the received language may be linked to the semantic analysis results. In such a way, a subsequent search may be performed on semantics, and the relevant conversations may be received. The process then proceeds to step 335 , where a decision is made whether to generate semantically related return language.
  • step 335 a decision is made whether to generate semantically related return language. If generating semantically related return language is desired, the Natural Language Generator 140 may generate language, at step 340 , that utilizes the semantics and knowledge representation from the Natural Language Analyzer 130 , from step 310 , along with cross referenced information from the Databases 110 . Said generated semantically related language may then be output to the Language Source 103 a to 103 z . In this manner return dialogue may be provided, enabling believable conversation.
  • step 345 a decision is made whether to store the generated semantically related language. If storing the generated language is desired, the generated language may be stored, or archived, in the Databases 110 at step 350 . Similarly to the analyzed language, stored generated language may be utilized for subsequent data mining and cross referencing. After storage of the generated language, the process ends.
  • step 345 if at step 345 , storing the generated language is not required, then the process ends.
  • step 335 if at step 335 , generating semantically related language is not desired, the process ends.
  • FIG. 4 shows a flow chart illustrating the process for analyzing language for knowledge representation and semantics shown generally at 310 .
  • the language is received.
  • the language is prepared. Preparation includes conversion, sentence identification and tokenization.
  • a decision is made whether the received language will be understood for semantics, or simply analyzed for knowledge representation.
  • the language is tagged at step 430 .
  • Part-of-Speech tagging occurs according to Table 1.
  • the part-of-speech tags database is utilized for tagging.
  • the text is chunked, or parsed, into the non-overlapping substructures of the sentence.
  • each sentence in the language is analyzed for grammatical mood.
  • the text is referenced across dictionaries, WordNet, and any applicable reference corpus.
  • the affect of the text is analyzed. Semantics are determined at step 480 by utilizing statistical pattern recognition.
  • knowledge representation is performed at step 490 .
  • additional language analyses may occur after step 490 , and before outputting in step 499 .
  • the illustration is intended to reflect this adaptability of language analysis.
  • the result of the natural language analysis is then output at step 499 .
  • step 490 knowledge representation is performed at step 490 .
  • additional language analyses may occur after step 490 , and before outputting in step 499 .
  • the illustration is intended to reflect this adaptability of language analysis.
  • the result of the natural language analysis is then output at step 499 .
  • Natural Language Processor 190 and methods thereof, disclosed within this application are very modular in nature. Where some embodiments including all of the stated components and steps are discussed, additional embodiments may exist that do not include all of the disclosed components and method steps. Similarly, the components and steps for the Natural Language Processor 190 listed are not intended to be an exhaustive list, as additional refinements become feasible and necessary.
  • FIG. 5 shows a flow chart illustrating the process for preparing language shown generally at 420 .
  • the beginning of this process is from step 410 , of FIG. 4 .
  • an inquiry is made as to if the language is in a text format. If the language is in an acceptable text format, the text sentences are separated, at step 523 . Then, at step 524 , the text is tokenized, wherein any special characters are removed and replaced with meaningful tokens. Tokenization cleans the text for downstream processing. Then the process ends by returning to step 430 , of FIG. 4 .
  • the language is converted into an acceptable text format at step 522 .
  • Text-to-text format conversions are relatively rapid and straightforward; however, it is intended in some embodiments of the invention, to also receive audio or waveform files and utilize speech recognition software in order to convert the language into text.
  • graphical input is intended to be received, such as sign language, body language and facial expressions.
  • a image recognition system may be utilized to convert graphical information into text for semantic analysis.
  • the text sentences are then separated, at step 523 .
  • the text is tokenized, wherein any special characters are removed and replaced with meaningful tokens. Then the process ends by returning to step 430 , of FIG. 4 .
  • FIG. 6 shows a flow chart illustrating the process for chunking text shown generally at 440 .
  • the beginning of this process is from step 430 , of FIG. 4 .
  • a placeholder N is set to a base value of 1.
  • the placeholder N is compared to the number of sentences in the text. If N is greater that the total sentences in the text, then the process ends by returning to step 450 of FIG. 4 . This ensures that all sentences within the corpus are chunked prior to the progression of the natural language processing.
  • a second placeholder M is set to 1 at step 613 . Then at step 614 the placeholder M is compared to the number of phrases in sentence N. If M is greater than the total phrases in the sentence N, then 1 is added to N in step 615 . This ensures that all phrases within sentence N are chunked prior to the chunking of subsequent sentences. The process then continues at step 612 , where the increased placeholder N is compared to the number of sentences in the text.
  • a query is made to determine if phrase M is a sentence at step 621 .
  • the rules utilized to chunk may be referred to above at Table 2. If the phrase M is a sentence, it is annotated as such at step 631 . Then at step 616 , 1 is added to M and the process then continues to step 614 , where the increased placeholder M is compared to the number of phrases in sentence N.
  • a query is made to determine if phrase M is a verb phrase at step 622 . If the phrase M is a verb phrase, it is annotated as such at step 632 . Then at step 616 , 1 is added to M and the process then continues to step 614 , where the increased placeholder M is compared to the number of phrases in sentence N.
  • a query is made to determine if phrase M is a noun phrase at step 623 . If the phrase M is a noun phrase, it is annotated as such at step 633 . Then at step 616 , 1 is added to M and the process then continues to step 614 , where the increased placeholder M is compared to the number of phrases in sentence N.
  • a query is made to determine if phrase M is a prepositional phrase at step 624 . If the phrase M is a prepositional phrase, it is annotated as such at step 634 . Then at step 616 , 1 is added to M and the process then continues to step 614 , where the increased placeholder M is compared to the number of phrases in sentence N.
  • phrase M is an adjective phrase at step 625 . If the phrase M is an adjective phrase, it is annotated as such at step 635 . Then at step 616 , 1 is added to M and the process then continues to step 614 , where the increased placeholder M is compared to the number of phrases in sentence N.
  • phrase M is a nominal phrase at step 626 . If the phrase M is a nominal phrase, it is annotated as such at step 636 . Then at step 616 , 1 is added to M and the process then continues to step 614 , where the increased placeholder M is compared to the number of phrases in sentence N.
  • a query is made to determine if phrase M is a gerund verb phrase at step 627 . If the phrase M is a gerund verb phrase, it is annotated as such at step 637 . Then at step 616 , 1 is added to M and the process then continues to step 614 , where the increased placeholder M is compared to the number of phrases in sentence N.
  • step 627 if, at step 627 the phrase is not a gerund verb phrase, 1 is added to M at step 616 .
  • the process then continues to step 614 , where the increased placeholder M is compared to the number of phrases in sentence N.
  • FIG. 7 shows a flow chart illustrating the process for setting grammatical mood shown generally at 450 .
  • the beginning of this process is from step 440 , of FIG. 4 .
  • a placeholder P is set to a base value of 1.
  • the placeholder P is compared to the number of sentences in the text. If P is greater that the total sentences in the text, then the process ends by returning to step 460 of FIG. 4 . This ensures that all sentences within the text are analyzed for grammatical mood prior to the progression of the natural language processing.
  • step 752 P is less than the total number of sentences within the text
  • the grammatical mood of sentence P is analyzed at step 753 .
  • step 754 a query is made whether sentence P is an interrogative sentence.
  • the definition for an interrogative sentence may be found at Table 3 above. Alternatively, a rule, statistical or hybrid identification system may be utilized as is discussed above. If the sentence is an interrogative sentence, the grammatical mood of sentence P is set to an interrogative structure at step 755 . Then, at step 759 , 1 is added to P and the process then continues to step 752 , where the increased placeholder P is compared to the total number of sentences within the text.
  • sentence P is not in an interrogative structure
  • a query is made whether sentence P is an imperative sentence at step 756 .
  • the definition for an imperative sentence may be found at Table 3 above. Alternatively, a rule, statistical or hybrid identification system may be utilized as is discussed above. If the sentence is an imperative sentence, the grammatical mood of sentence P is set to an imperative structure at step 757 . Then, at step 759 , 1 is added to P and the process then continues to step 752 , where the increased placeholder P is compared to the total number of sentences within the text.
  • step 756 additional grammatical mood inquiries may occur after step 756 , and before setting the sentence mood to declarative in step 758 .
  • other mood types may be incorporated within some embodiments. The illustration is intended to reflect this adaptability of grammatical mood analysis.
  • sentence P is not in an imperative structure
  • the grammatical mood of sentence P is set to a declarative structure at step 758 .
  • This provides that the declarative structure is default for both true declarative sentences, and in the case of sentences whose grammatical mood is difficult to ascertain.
  • 1 is added to P and the process then continues to step 752 , where the increased placeholder P is compared to the total number of sentences within the text.
  • FIG. 8 shows a flow chart illustrating the process for pattern recognition shown generally at 480 .
  • the beginning of this process is from step 470 , of FIG. 4 .
  • synonyms are identified for the elements of the text. Synonyms are different lexemes with the same meaning. Often, words are not synonyms for each other in any context but they are context bound.
  • homonyms are identified for the elements of the text. Homonyms are words of the same orthographic and/or phonographic form with unrelated meaning e.g. mouse (computer input device) and mouse (animal). This type of relationship is typically, from a semantic perspective, of little interest.
  • hyponyms are identified for the elements of the text.
  • Hyponyms are the more specific lexeme of a pair of lexemes, when one lexeme denotes a subclass of the other.
  • a ‘human’ is a hyponym of ‘primate’.
  • hypernyms are identified for the elements of the text. Hypernyms are the less specific lexeme of a pair of lexemes, when one lexeme denotes a subclass of the other.
  • a ‘primate’ is a hypernym of ‘human’.
  • Synonyms, hyponyms, hypernyms and homonyms are identified through the referencing of the elements through dictionaries and thesauruses as well as statistical pattern recognition.
  • step 884 additional semantic analyses may occur after step 884 , and before lexical chain generation in step 889 .
  • the illustration is intended to reflect this adaptability of semantic analysis.
  • lexical chains are generated for the text.
  • lexical chains are data-structures that represent the semantics of a given text at a very abstract level.
  • a lexical chain is made up of a set of different nouns and/or verbs that represent the semantics of a given text. They occur in any order.
  • a text may usually be represented by more than one lexical chain.
  • FIG. 9 shows a flow chart illustrating the process for outputting post operations shown generally at 499 .
  • the beginning of this process is from step 490 , of FIG. 4 .
  • a query is made whether data mining is to be performed upon the semantically analyzed language. If data mining is desired, a data report is generated, at step 990 , wherein trends, semantics, ad targets and product placement suggestions are compiled for Data Miner 120 consumption.
  • the generated data report may be personalized to the Language Source 103 a to 103 z providing the dialogue, or language.
  • the semantic analysis data report is output to the Data Miner 120 , in step 991 . Then the process proceeds to step 902 .
  • step 902 a query is made as to whether a response is required.
  • Grammatical mood plays an important role in the determination of when responses are required. If a response is required, the semantic analysis is output to the Natural Language Generator 140 in step 992 . Then the process proceeds to step 903 .
  • step 903 a query is made as to whether persistent knowledge of the conversation is required.
  • the type of avatar, content of conversation and length of conversation may all play a role in the persistence of the dialogue. For example, in a game scenario it may be advantageous for an avatar to “remember” lengthy or involved conversation, but “forget” smaller and insignificant dialogue. Not only does this provide the avatar with a believable behavior scheme, but less data-rich content (for data mining purposes) may be eliminated, thus reducing the resources required for storage.
  • the semantic analysis is formatted for retrieval in step 993 .
  • step 994 the formatted semantic analysis is stored within the conversation history database. The process then ends. Else if persistent knowledge is not desired, at step 903 , the process ends.
  • FIG. 10 shows a flow chart illustrating the process for formatting processed text for retrieval shown generally at 993 .
  • the beginning of this process is from step 903 , of FIG. 9 .
  • the processed sentences are separated.
  • the sentences are chunked by a hash table, wherein the hash table includes the Part-of-Speech tags and words, thus allowing for rapid retrieval of the processed conversation. Additionally, stored conversation may be referenced by the conversant.
  • hash table formatting the process ends by proceeding to step 994 of FIG. 9 .
  • FIGS. 11 to 16 disclose specific implementations of the Natural Language Processor 190 discussed above in FIGS. 1 to 10 . While the methods disclosed in FIGS. 11 to 16 illustrate specific implementations of some embodiments, these figures and descriptions are intended to be exemplary in nature.
  • FIG. 11 shows a flow chart illustrating the process for an embodiment of data preparation for semantic searches, shown generally at 1100 , utilizing the Natural Language Processor 190 of the Autonomous Avatar Driver System 100 .
  • the illustrated method in FIG. 11 is intended to be useful in conjunction with a Frequently Asked Question (FAQ) style data set.
  • a FAQ typically includes a set of predefined questions, followed by an answer to said question.
  • this embodiment of the semantic searching has particularly important implications for corporate customers and guide avatars.
  • a corporation could, in some embodiments, replace its existing FAQ page on its website with a helper avatar.
  • the avatar may be graphically represented, but may often have no physical embodiment. In such embodiments, there may be a field where the inquirer, the user, may type her question.
  • Such an embodiment may mimic a chat style conversation that currently occurs with a customer representative.
  • the avatar may then perform the semantic search of the FAQ database to return the question-answer pairs that are the most relevant to the inquiry.
  • Current help menus provide searching for troubleshooting purposes, however these searches are greatly limited in their ability to comprehend semantics, and inability to process natural language dialogue.
  • these hurdles may be minor due to the proficient usage of search terms.
  • many corporations cater to less technically sophisticated clientele, and the ability for these companies to outreach effectively to their customers may be greatly enhanced by allowing natural language searches that often provide better search results than a traditional troubleshooting search; even one performed by a sophisticated and proficient searcher.
  • guide characters, or tutorial characters may be desired. In these circumstances, it is of particular importance to be able to interpret natural language and provide meaningful responses for the avatar to be believable, and to enhance the experience for the user.
  • the process begins at step 1101 , where the answer data is queried.
  • the answer data is the answer portion of a FAQ question-answer pair.
  • Each answer is queried, and in step 523 sentence detection is performed upon the answer data.
  • the answer data is tokenized to clean up the answer data by removing all special character and replacing them with meaningful tokens.
  • the parts of speech are tagged in step 430 , utilizing, in some embodiments, the tags defined in Table 1.
  • the answer data is then chunked, in step 440 , in a manner similar to that previously discussed, and in reference to Table 2. After chunking the process diverges from the prototypical method of natural language processing disclosed earlier in the present invention.
  • the nouns and verbs of the answer data are extracted.
  • the question portion of the FAQ question-answer pair is inputted at step 1107 .
  • the semantic similarity between the noun and verbs extracted from the answer, and the question is then determined in step 1108 .
  • Semantics of the answers is performed by determining the semantics of the nouns and verbs extracted through cross referencing the words and through pattern recognition. Similarly, the semantics of the inputted question are determined through cross referencing the words and pattern recognition. The overlap between the semantic analysis of the question, and semantic analysis of the noun-verb list of the answer, results in the semantic similarity.
  • the semantic similarity is then utilized in 1109 to generate a lexical chain.
  • step 1110 the generated lexical chain is stored in reference to the question-answer pair it originated from. This process is, in some embodiments, performed for all question-answer pairs. Moreover, this procedure is only required to be completed once, and repeated only whenever new data is introduced in the FAQ. It should be noted that there are many methods of preparing a lexical chain for a question-answer pair in a FAQ; however, due to the semantically similar relationship between the question and its answer the abovementioned process develops a finely tuned lexical chain that is highly relevant to the core meanings of the FAQ pair. By storing these lexical chains, referenced back to their original question-answer pairs, subsequent inquiries may leverage these chains to provide meaningful semantic searches, as may be seen in FIG. 12 .
  • FIG. 12 shows a flow chart illustrating the process for an embodiment of semantic searches for frequently-asked-questions, shown generally at 1200 . This process follows the process illustrated generally at 1100 at FIG. 11 .
  • Interactive question refers to a question originating from an end-user of the avatar. This end-user may be the corporation's customer accessing their FAQ on the website, or may be a gamer interacting with a guide avatar. These interactive questions do not, in these embodiments, refer to the question portions of the FAQ.
  • the interactive question is processed for semantics and knowledge representation by the Natural Language Analyzer 130 through a process similar to that disclosed at FIG. 4 .
  • the results of the semantic analysis of the question are then compared to the lexical chains generated in step 1109 of FIG. 11 for semantic similarity at step 1203 .
  • the lexical chains with the greatest semantic similarity to the semantics of the interactive question may then be referenced to the question-answer pair that gave birth to it.
  • These question-answer pairs may be displayed to the end user. Alternatively, only the answer portion of the question-answer pairs may be displayed to the end user.
  • the answer may be displayed after being reprocessed through the Natural Language Analyzer 130 and the Natural Language Generator 140 in order to provide an answer that is uniquely tailored to the interactive question, and that may include knowledge representation.
  • a guide character traditionally avoided by gamers due to its daunting monotony and longwinded rigidity, may provide a continuation of the game narrative that is believable and enjoyable to the gamer.
  • FIG. 13A shows a flow chart illustrating the process for one embodiment of determining semantic similarities for semantic searches shown generally at 1203 A.
  • the beginning of this process is intended to be from step 1202 , of FIG. 12 .
  • this process 1203 A is not limited to such context, and may be employed in many suitable semantic similarity comparisons.
  • a placeholder T is set to a base value of 1.
  • the placeholder T is compared to the number of lexical chains that were generated in step 1109 of FIG. 11 for the question-answer pairs of the FAQ. If T is greater that the total number of lexical chains, then the process proceeds to step 1311 where lexical chains are ranked by total weights. This ensures that all lexical chains generated for the FAQ are compared prior to the ranking of semantic similarities.
  • a second placeholder U is set to 1 at step 1303 .
  • the placeholder U is compared to the number of elements in lexical chain T. If U is greater than the total elements in the lexical chain T, then the weight of the lexical chain T is multiplied by a modifier Z T at step 1308 .
  • Modifier Z T provides for corrections to be made for differences between lexical chains, such as for chain length. For example, a highly refined question-answer pair will tend to produce a relatively short lexical chain, since the subject matter explored by the FAQ entry is refined and specific. However, a broad question and answer pair may result in a very long lexical chain.
  • the interactive question may have more aggregate similarities with the long lexical chain due to its broad nature; however the interactive question is better answered by the narrower, in-depth FAQ entry.
  • the modifier Z T may be applied to increase the weight of the shorter lexical chain's semantic similarities.
  • step 1309 a value of 1 is added to T in step 1309 . This ensures that all elements within the lexical chain T are compared to the semantics of the interactive question. The process then returns to step 1302 , where the increased placeholder T is compared to the number of lexical chains for the FAQ.
  • a weight is assigned to element U by comparing the semantics of the interactive question to element U. A direct semantic match may generate a higher weight value to element U than a more attenuated semantic relationship.
  • the weight assigned to element U is added to the total weight of lexical chain T.
  • a value of 1 is added to U.
  • the process then returns to step 1304 , where the increased placeholder U is compared to the number of elements of the lexical chain T. In this manner the total weight of the lexical chain T is the sum of the weights of its respective elements.
  • the lexical chains may be ranked by said total weights in step 1311 .
  • the lexical chains with the largest total weights are referenced back to their parent question-answer pairs, and the question-answer pair is outputted for display or processing. The process then ends. In this way, the interactive question returns the FAQ entry or entries that are most relevant to the inquiry.
  • FIG. 13B shows a flow chart illustrating the process for another embodiment of determining semantic similarities for semantic searches shown generally at 1303 B.
  • This process is similar to the process outlined in FIG. 13A , but requires less computational resources and may produce minor variation in results. The beginning of this process is intended to be from step 1202 , of FIG. 12 . However, this process 1203 B is not limited to such context, and may be employed in many suitable semantic similarity comparisons.
  • a placeholder T is set to a base value of 1.
  • the placeholder T is compared to the number of lexical chains that were generated in step 1109 of FIG. 11 for the question-answer pairs of the FAQ. If T is greater that the total number of lexical chains, then the process proceeds to step 1329 where lexical chains are ranked by total weights. This ensures that all lexical chains generated for the FAQ are compared prior to the ranking of semantic similarities.
  • a second placeholder U is set to 1 at step 1322 . Then at step 1323 the placeholder U is compared to the number of elements in lexical chain T. If U is greater than the total elements in the lexical chain T then the weight of the lexical chain T is multiplied by a modifier Z T at step 1327 .
  • Modifier Z T provides for corrections to be made for differences between lexical chains, such as for chain length. For example, a highly refined question-answer pair will tend to produce a relatively short lexical chain, since the subject matter explored by the FAQ entry is refined and specific. However, a broad question and answer pair may result in a very long lexical chain.
  • the interactive question may have more aggregate similarities with the long lexical chain due to its broad nature; however the interactive question is better answered by the narrower, in-depth FAQ entry.
  • the modifier Z T may be applied to increase the weight of the shorter lexical chain's semantic similarities.
  • step 1328 a value of 1 is added to T in step 1328 . This ensures that all elements within the lexical chain T are compared to the interactive question. The process then returns to step 1321 , where the increased placeholder T is compared to the number of lexical chains for the FAQ.
  • step 1323 if, at step 1323 , U is less than the total elements in the lexical chain T, then, at step 1324 , a query is made if the element U of lexical chain T is matched by any element in the interactive question. If element U matches an element in the question, then 1 is added to the total weight of the lexical chain T in step 1325 . Then, in step 1326 , a value of 1 is added to U. This ensures that all elements U within the lexical chain T are compared to the interactive question. The process then returns to step 1323 , where the increased placeholder U is compared to the number of elements of the lexical chain T.
  • step 1324 element U does not match any element in the question, then a value of 1 is added to U at step 1326 . This ensures that all elements U within the lexical chain T are compared to the interactive question. The process then returns to step 1323 , where the increased placeholder U is compared to the number of elements of the lexical chain T. In this manner the total weight of the lexical chain T is the sum of the direct matches of its elements and the question.
  • the lexical chains may be ranked by said total weights in step 1329 .
  • the lexical chains with the largest total weights are referenced back to their parent question-answer pairs, and the question-answer pair is outputted for display or processing. The process then ends. In this way, the interactive question returns the FAQ entry or entries that are most relevant to the inquiry.
  • FIG. 14 shows a flow chart illustrating the process for an embodiment of a general semantic search utilizing the natural language processor system, shown generally at 1400 .
  • Process 1400 is a more general semantic searching method then those embodiments disclosed in FIGS. 11 to 13B .
  • process 1400 does not require a FAQ style data set, but rather answer data in the form of a plurality of answer entries. These answer entries may include separate data sources, previous dialogue conversations, news articles, press releases, or virtually any series of informational text.
  • Step 1401 and 1402 The process begins by parallel process of Step 1401 and 1402 . Either step 1401 or 1402 may be performed first, or they may be performed simultaneously.
  • a question language is received.
  • the language question is processed through natural language analysis as illustrated in FIG. 4 .
  • answer entries are queried. Said answer entries are analyzed through the Natural Language Analyzer 130 for semantics at step 310 .
  • the semantics for each answer entry is compared for relatedness to the results the semantics for the question determined at steps 310 .
  • step 1408 the results of the relatedness between the semantics of each answer entry and the question are outputted for further processing. The process then ends.
  • FIG. 15A shows a flow chart illustrating the process for one embodiment of determining semantic similarities for general semantic searches utilizing the Autonomous Avatar Driver System 100 shown generally at 1407 A.
  • the beginning of this process is intended to be from step 1406 , of FIG. 14 .
  • this process 1407 A is not limited to such context, and may be employed in any suitable semantic similarity comparison.
  • a placeholder Q is set to a base value of 1.
  • the placeholder Q is compared to the number of answer entries in the answer data set. If Q is greater that the total number of answer entries then the process proceeds to step 1510 where answer entries are ranked by total weights. This ensures that all answer entries are compared prior to the ranking of semantic similarities.
  • a second placeholder R is set to 1 at step 1503 . Then at step 1504 the placeholder R is compared to the number of elements in lexical chain of answer entry Q. If R is greater than the total elements in the lexical chain of answer entry Q then the weight of the lexical chain of answer entry Q is multiplied by a modifier Z Q at step 1508 .
  • Modifier Z Q provides for corrections to be made for differences between lexical chains, such as for chain length. For example, a highly refined answer entry will tend to produce a relatively short lexical chain. However, a broad answer entry may result in a very long lexical chain.
  • the interactive question may have more aggregate similarities with the long lexical chain due to its broad nature; however the interactive question is better answered by the narrower, in-depth answer entry.
  • the modifier Z Q may be applied to increase the weight of the shorter lexical chain's semantic similarities.
  • step 1509 a value of 1 is added to Q in step 1509 . This ensures that all elements within the lexical chain of answer entry Q are compared to the semantics of the question. The process then returns to step 1502 , where the increased placeholder Q is compared to the number of answer entries.
  • a weight is assigned to element R by comparing the semantics of the question to element R. A direct semantic match may generate a higher weight value to element R than a more attenuated semantic relationship.
  • the weight assigned to element R is added to the total weight of answer entry Q.
  • a value of 1 is added to R. The process then returns to step 1504 , where the increased placeholder R is compared to the number of elements of the lexical chain of answer entry Q. In this manner the total weight of the lexical chain of answer entry Q is the sum of the weights of its respective elements.
  • the answer entries may be ranked by said total weights in step 1510 .
  • the process then ends by proceeding to step 1408 in FIG. 14 .
  • FIG. 15B shows a flow chart illustrating the process for another embodiment of determining semantic similarities for general semantic searches utilizing the Autonomous Avatar Driver System 100 , shown generally at 1407 B.
  • This process is similar to the process outlined in FIG. 15A , but requires less computational resources and may produce minor variation in results. The beginning of this process is intended to be from step 1406 , of FIG. 14 . However, this process 1407 B is not limited to such context, and may be employed in any suitable semantic similarity comparison.
  • a placeholder Q is set to a base value of 1.
  • the placeholder Q is compared to the number of answer entries. If Q is greater that the total number of answer entries then the process proceeds to step 1530 where answer entries are ranked by total weights. This ensures that all answer entries are compared prior to the ranking of semantic similarities.
  • a second placeholder R is set to 1 at step 1523 . Then at step 1524 the placeholder R is compared to the number of elements in lexical chain of answer entry Q. If R is greater than the total elements in the lexical chain of answer entry Q then the weight of the lexical chain of answer entry Q is multiplied by a modifier Z Q at step 1528 .
  • Modifier Z Q provides for corrections to be made for differences between lexical chains, such as for chain length. For example, a highly refined answer entry will tend to produce a relatively short lexical chain. However, a broad answer entry may result in a very long lexical chain.
  • the question may have more aggregate similarities with the long lexical chain due to its broad nature; however the interactive question is better answered by the narrower, in-depth answer entry.
  • the modifier Z Q may be applied to increase the weight of the shorter lexical chain's semantic similarities.
  • step 1529 a value of 1 is added to Q in step 1529 . This ensures that all elements within the lexical chain of answer entry Q are compared to the question. The process then returns to step 1522 , where the increased placeholder Q is compared to the number of answer entries.
  • step 1524 if, at step 1524 , R is less than the total elements in the lexical chain of answer entry Q, then, at step 1525 , a query is made if the element R of lexical chain of answer entry Q is matched by any element in the question. If element R matches an element in the question, then 1 is added to the total weight of answer entry Q in step 1526 . Then, in step 1527 , a value of 1 is added to R. This ensures that all elements R within the lexical chain of answer entry Q are compared to the question. The process then returns to step 1524 , where the increased placeholder R is compared to the number of elements of the lexical chain of answer entry Q.
  • step 1525 element R does not match any element in the question, then a value of 1 is added to R at step 1527 . This ensures that all elements R within the lexical chain of answer entry Q are compared to the question.
  • the process then returns to step 1524 , where the increased placeholder R is compared to the number of elements of the lexical chain of answer entry Q. In this manner the total weight of answer entry Q is the sum of the direct matches of its lexical chain's elements and the question.
  • the answer entries may be ranked by said total weights in step 1530 .
  • the process then ends by proceeding to step 1408 in FIG. 14 .
  • FIG. 16 shows a flow chart illustrating the process for generating tag clouds shown generally at 1600 .
  • Tag clouds are a collection of keywords that may be used in order to enable the user to interact more efficiently with a user interface.
  • Tag clouds are well known by those skilled in the art; however, the usage of tag clouds that are responsive to natural language processing is unique to the present invention.
  • This embodiment of the present invention deviates from the semantic searching embodiments; however, this embodiment is intended to illustrate another useful application of the Autonomous Avatar Driver System 100 .
  • the process begins at step 1601 , where conversation text data is queried for downstream semantic processing. Beyond conversation data, additional data may be available for query, such as conversation histories, databases and external data sources. In this way the avatar may be able to generate tag clouds that provide somewhat impulsiveness conversation developments. Such impulsiveness allows more avenues of conversation to develop, thereby enhancing the users' experience. Additionally, impulsiveness is a decidedly human trait, thus adding to the believability of the avatar. For example, in a gaming context if the avatar is assigned a role as a farmer, and the user engages in language regarding the weather, the tag cloud may provide keywords about weather, weather predictions and weather's effects upon crops. This is made possible by supplying the avatar with knowledge templates regarding farming practices, as is discussed previously.
  • sentence detection is performed on the queried text.
  • the data is tokenized, thereby cleaning the text by removing special characters.
  • the cleansed text is tagged as is discusses above at Table 1.
  • the tagged text is then parsed, or chunked, in step 440 , as was seen at Table 2.
  • the nouns and verbs are extracted from the text in order to generate lexical chains at step 1607 .
  • the lexical chains are then filtered at step 1608 to generate tag clouds.
  • the tag clouds are generated.
  • the filtering may filter individual chains, or filter multiple chains. For instance, the available lexical chains may be compared for similar content. Lexical chains that are too similar may be synthesized, or one of them may be eliminated. This reduces the redundancy of lexical chains used in the construction of the tag clouds. For individual lexical chains, it may be advantageous to reduce each chain to a predetermined number of elements. Redundant elements may be synthesized or eliminated first. Among strings of hyponyms and hypernyms an element with a medium level of detail may be retained and the remaining hyponyms and hypemyms may be discarded. Again, the purpose of such filtering is the reduction of the redundant elements used in the construction of the tag clouds.
  • the generation of the tag clouds may be as simple as the displaying of the remaining elements of the lexical chains. Alternatively, intact chains may be displayed. Moreover, in some embodiments, the tag clouds could even include dialog from the Natural Language Generator 140 , which may utilize the lexical chains generated in step 1607 to generate natural language “thoughts” or conversation suggestions in natural language.
  • FIG. 17 shows a schematic block diagram illustrating the Autonomous Avatar Driver 150 .
  • the Autonomous Avatar Driver 150 includes a Sourcer 1711 , a Segmenter 1712 , a Parser 1713 , a Semantic Engine 1714 , an Analytics Engine 1715 , a Pragmatics Analyzer 1716 , a Recommender 1717 and a Response Generator 1718 .
  • Each component of the Autonomous Avatar Driver 150 may be coupled with one another, thereby enabling communication between the respective components of the Autonomous Avatar Driver 150 . It should be noted that additional, or fewer, components may be included in the Autonomous Avatar Driver 150 as is desired.
  • the illustrated Autonomous Avatar Driver 150 is intended to be purely exemplary in nature. In some embodiments, the Autonomous Avatar Driver 150 may utilize the Natural Language Processor 190 for many of the analytic and semantic analysis of the avatar dialog.
  • the exemplary Data Source 102 a and Language Source 103 a may provide input to the Autonomous Avatar Driver 150 via the Sourcer 1711 .
  • the Segmenter 1712 may segment the language, and the Parser 1713 may parse the language.
  • the Semantic Engine 1714 may receive the segmented and parsed language and perform a semantic analysis. Details of these processes, as performed for these particular embodiments of the Autonomous Avatar Driver 150 are further explained below. Of course alternate methods of segmenting, parsing and semantic analysis of the received language may be utilized.
  • the generic Natural Language Processor 190 may segment, parse and analyze for the received language semantics as previously disclosed.
  • the Analytics Engine 1715 performs analytic analysis to the semantically analyzed language.
  • Analytics may include filtering for common words, token frame statistics and N-gram analysis.
  • the Pragmatics Analyzer 1716 may take the language and resolve pragmatic issues, such as slang usage, knowledge template application, grouping of proper nouns, and estimation of affect.
  • the Recommender 1717 provides recommendations from the analyzed language to the Data Consumer 125 . These recommendations may include highly relevant contextual advertising, statistical analysis, and improvement of the user's experience through user recommendations. For example, if the dialog is regarding music, the analyzed language could identify the user's preference of bands. These preferences may then be compared to statistical databases of similar user's preferences to provide music that is likely to be enjoyed by the user. Such “product recommendation” is not new by itself, however by intelligently and contextually analyzing a dialog the Autonomous Avatar Driver 150 may provide highly relevant, accurate and useful recommendations.
  • the Recommender 1717 may provide recommendations within a social network environment of other individuals that have similar interests, and/or compatible personalities. This may be of particular use in casual social networks such as Facebook®, as well dating style social networks such as match.com®.
  • the Response Generator 1718 may utilize the analyzed language to generate relevant responses that may in turn drive the Autonomous Avatar 1720 .
  • Such response information may include graphical information, such as avatar expression, as well as relevant dialog.
  • the Response Generator 1718 may, additionally, receive recommendation information from the Recommender 1717 to incorporate the generated recommendation within the response. Therefore, new and relevant topics, tailored to the conversation, may be included in the response. Additionally, Autonomous Avatar 1720 may provide relevant and contextual advertising when appropriate.
  • FIG. 18 shows a schematic block diagram illustrating the Sourcer 1711 of the Autonomous Avatar Driver 150 .
  • the Sourcer 1711 may include a Dialog Manager 1802 and a Source Crawler 1804 .
  • the Dialog Manager 1802 may receive dialog information from the Language Source 103 p .
  • the Sourcer 1711 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150 .
  • the Source Crawler 1804 may, in some embodiments, provide a more active role of crawling the Data Source 102 a for information as pertains to the dialog received by the Dialog Manager 1802 .
  • the Dialog Manager 1802 may be coupled to the Source Crawler 1804 to facilitate the searching for relevant information by the Source Crawler 1804 .
  • the Source Crawler 1804 may utilize any crawling technology, as is well known by those skilled in the art.
  • the Source Crawler 1804 may search any appropriate Data Source 102 a ; however, social media, and structured data sources, such as Wikipedia, may be routinely sourced due to the ease of searching, as well as the utility of the data gathered for most interactions. For example, the format of typical web-logs, or blogs, is very similar to a screenplay. Screenplays and blogs typically have topical formats, which are highly useful for gathering additional contextual information.
  • FIG. 19 shows a schematic block diagram illustrating the Segmenter 1712 of the Autonomous Avatar Driver 150 .
  • the Segmenter 1712 includes a Particle Representor 1902 , a Character Converter 1904 and a Splitter 1906 , each couple to one another.
  • the Segmenter 1712 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150 .
  • the Particle Representor 1902 may separate the dialog into “particles”, or discrete logical subsections of the dialog. For example, “particles” of a screenplay would include the individual acts and scenes. This particle representation may then represent the language in a data structure format. Particle representation may additionally include speaker detection.
  • the Character Converter 1904 may convert the dialog characters. In some embodiments, the Character Converter 1904 will convert the dialog to UTF-8 encoding. However, additional conversion formats may be utilized by the Character Converter 1904 as is known by those skilled in the art.
  • the Splitter 1906 may split the sentences of the dialog.
  • a max entropy splitter may be utilized for the Splitter 1906 .
  • FIG. 20 shows a schematic block diagram illustrating the Parser 1713 of the Autonomous Avatar Driver 150 .
  • the Parser 1713 may include a Grammar Linker 2002 , a Grammatical Mood Analyzer 2004 , a Tagger 2006 and a Variant Pruner 2008 , each coupled to one another.
  • the Parser 1713 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150 .
  • the Grammar Linker 2002 may link grammatically useful word combinations. These combinations provide precision of the language when transformed into computer language. An example of words that may be linked includes “have” and “being”. These word combinations denote specific grammatical stances when paired and conjugated.
  • the Grammatical Mood Analyzer 2004 may determine the grammatical mood of the language.
  • the grammatical moods include interrogative, declarative and imperative. See table 3 above. Of course additional moods may be determined, such as energetic and conditional, as is desired for functionality.
  • the Tagger 2006 may tag the words of the dialog, as is discussed above in the natural language processor section. Tagging the language is fundamental to the generation of lexical chains, as well as the ability for the Autonomous Avatar Driver 150 to derive accurate semantic information from the language. In some embodiments, a max entropy tagger may be utilized for the Tagger 2006 .
  • the Variant Pruner 2008 may prune variants within the dialog by utilizing Bayesian statistics. Results from the Variant Pruner 2008 may be output to the NLP Servicer 2020 .
  • the NLP Servicer 2020 may include any entity that requires the parsed language. Examples of the NLP Servicer 2020 include chat rooms or other data compilers.
  • FIG. 21 shows a schematic block diagram illustrating the Semantic Engine 1714 of the Autonomous Avatar Driver 150 .
  • the Semantic Engine 1714 may include a Token Framer 2102 , a Lemmatizer 2104 , a Semantic Network Generator 2106 and a Co-reference Resolver 2108 , each coupled to one another.
  • the Semantic Engine 1714 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150 .
  • the Token Framer 2102 may tokenize the dialog as discussed previously.
  • the Lemmatizer 2104 may lemmatize the dialog, wherein the verbs and nouns are reduced back into basic forms.
  • the Semantic Network Generator 2106 may generate meaningful graphs of the concepts, wherein relationships of the concepts are defined. Statistical profiling may be utilized in the generation of the semantic networks.
  • Such semantic networks may require significant processing power and time to complete.
  • corpuses may be utilized to pre-compute semantic networks, which may be updated as required. These pre-computed semantic networks may then be referenced when the analyzed language contains nodes of commonality. Thus processing time may be reduced for accurate semantic network generation.
  • the Co-reference Resolver 2108 may resolve co-referencing ambiguities within the dialog.
  • the Co-reference Resolver 2108 may resolve the co-references via pattern matching, context appropriateness and by utilizing verb reference to a subject.
  • FIG. 22 shows a schematic block diagram illustrating the Analytics Engine 1715 of the Autonomous Avatar Driver 150 .
  • the Analytics Engine 1715 may include a Common Word Filter 2202 , a Token Frame Statistics Analyzer 2204 , and an N-gram Analyzer 2206 , each coupled to one another.
  • the Analytics Engine 1715 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150 .
  • the Common Word Filter 2202 may filter the dialog for common words, thereby reducing redundancy.
  • the Token Frame Statistics Analyzer 2204 may sum lemmatized dialog. Such a count is akin to “keyword hits” utilized by standard searching systems; however by counting the lemmatized dialog, more contextually accurate searches are possible. Additionally, the Token Frame Statistics Analyzer 2204 may also perform additional statistical analysis on the dialog as is desired for system functionality. An example would include nodal analysis of the semantic network.
  • the N-gram Analyzer 2206 may perform an N-gram analysis on the dialog.
  • typically 3 or 4 words may be utilized for the N-gram analysis.
  • different word numbers may be utilized as is desired.
  • the number of words utilized for the N-gram analysis is highly language dependent.
  • a sliding window along N words may be utilized to generate popularity statistics.
  • FIG. 23 shows a schematic block diagram illustrating the Pragmatics Analyzer 1716 of the Autonomous Avatar Driver 150 .
  • the Pragmatics Analyzer 1716 may include a Slang Resolver 2302 , a Knowledge Template Resolver 2304 , a Proper Noun Grouper 2306 and an Affect Estimator 2308 , each coupled to one another.
  • the Pragmatics Analyzer 1716 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150 .
  • the Pragmatics Analyzer 1716 enables the Autonomous Avatar Driver 150 to filter certain behaviors of human users that make contextual analysis of the dialog problematic.
  • the Slang Resolver 2302 may resolve slang language utilizing statistical collection of word usage as gained from ongoing corpuses, such as blogs. These slang terms may then populate a slang database for rapid retrieval when resolving slang in the dialog.
  • the Knowledge Template Resolver 2304 may provide knowledge representation via knowledge templates as disclosed above in the discussion of the natural language processing.
  • the Proper Noun Grouper 2306 may group together proper nouns as identified by grammatical form.
  • the Affect Estimator 2308 may estimate dialog affect, or emotion, by analyzing the lemmatized dialog in conjunction with the grammatical mood of the dialog.
  • FIG. 24 shows a schematic block diagram illustrating the Recommender 1717 of the Autonomous Avatar Driver 150 .
  • the Recommender 1717 may include a Cloud Aggregator 2402 , a Hood Clusteror 2404 , a Social Network Recommender 2406 and a Contextual Advertising Generator 2408 , each coupled to one another.
  • the Recommender 1717 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150 .
  • Each component of the Recommender 1717 may output their recommendations to the Data Consumer 125 .
  • the Cloud Aggregator 2402 may aggregate the tag clouds of related dialog.
  • the Hood Clusteror 2404 may cluster people by their closest distance by interest or compatibility. This may have particular utility for “people fit” services on social networks and dating services.
  • the Social Network Recommender 2406 may recommend contacts on social networks. This may be of particular use for business to business contact generation.
  • the Contextual Advertising Generator 2408 may generate highly contextual and highly targeted advertising. Such advertising recommendations are only possible by analyzing not only content, but also meaning of the dialog.
  • the Recommender 1717 may assist the Response Generator 1718 to provide dialog for the Autonomous Avatar 1720 that is not only responsive, but also that may include recommendations that are relevant to the user.
  • FIG. 25 shows a schematic block diagram illustrating the Response Generator 1718 of the Autonomous Avatar Driver 150 .
  • the Response Generator 1718 may include a Summarizer 2502 and a Dialog Manager 2504 coupled to one another.
  • the Response Generator 1718 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150 .
  • the Summarizer 2502 may summarize the response that is generated.
  • the Dialog Manager 2504 may then manage the dialog for the Autonomous Avatar 1720 . It should be noted that while the Response Generator 1718 is discussed as generating dialog, additional responses may be generated by the Response Generator 1718 , including but not limited to graphical emotional expressions, movements and non-word exclamations.
  • FIG. 26A shows a schematic block diagram of the Virtual Universe 2600 for the autonomous avatar driver system of FIG. 17 .
  • the Virtual Universe 2600 may be broken down into five subcategories: Virtual Overlays of Real World Data 2601 , WEB communities 2602 , Massively Multiplayer Online Games (MMOGs) 2603 , Social Worlds 2604 , and Telecom 2605 .
  • Virtual Overlays of Real World Data 2601 include, but are not limited to, Google Earth and Microsoft Flight Simulator X.
  • Examples of WEB communities 2602 include, but are not limited to, YouTube and MySpace.
  • Examples of MMOGs 2603 include, but are not limited to, World of Warcraft, Guild Wars and Hive.
  • Examples of Social Worlds 2604 include, but are not limited to, Second Life and Neopets.
  • Telecom 2605 examples include, but are not limited to, cell phones, BlackBerry Devices and Personal Digital Assistants (PDA). Additional subcategories may exist, or may emerge with new technology. It is intended that these additional subcategories be incorporated into the Virtual Universe 2600 .
  • the Autonomous Avatar Driver 150 is coupled to the subcategories of the Virtual Universe 2600 through the WAN 101 .
  • FIG. 26B A logical block diagram of the Virtual Universe 2600 is shown in FIG. 26B .
  • Each Virtual Environments 2611 a , 2611 b to 2611 x , 2612 a , 2612 b to 2612 y , 2613 a , 2613 b to 2613 z , 2614 a , 2614 b to 2614 m , 2615 a , 2615 b to 2615 n is coupled to the WAN 101 .
  • Each subcategory, Virtual Overlays of Real World Data 2601 , WEB Communities 2602 , MMOGs 2603 , and Social Worlds 2604 , and Telecom 2605 may include multiple Virtual Environments 2611 a to 2615 n .
  • Virtual Environments 2611 a to 2615 n may be hybrids of these subcategories.
  • the boundaries between individual Virtual Environments 2611 a to 2615 n are distinct and nearly impassable.
  • the Virtual Overlays of Real World Data 2601 have provided some connectivity between Virtual Environments 2611 a to 2615 n as shown in FIG. 25 ; however this connectivity is limited in scope.
  • the Autonomous Avatar Driver 150 is able to access all the Virtual Environments 2611 a to 2615 n thereby enabling the driving of the Autonomous Avatar 1720 in any Virtual Environment 2611 b.
  • FIG. 26C A logical block diagram of an exemplary Virtual Environment 2611 b is shown in FIG. 26C .
  • each Virtual Environment 2611 b exists an Enabler 2631 .
  • the Enabler 2631 allows for Autonomous Avatar 1720 a , 1720 b to 1720 t to access the WAN 101 , and eventually the Autonomous Avatar Driver 150 .
  • each Virtual Environment 2611 b has a corresponding Enabler 2631 .
  • any number of Autonomous Avatars 1720 a to 1720 t may exist within a Virtual Environment 2611 b at any given time.
  • Autonomous Avatar 1720 a may include Physical Attributes 2641 , Intellectual Attributes 2642 and Emotional Attributes 2643 . Each attribute may be coupled to an Avatar Engine 2644 which may coordinate and manage each attribute.
  • Physical Attributes 2641 may include Autonomous Avatar's 1720 a physical statistics, such as strength, and appearance data.
  • Intellectual Attributes 2642 may include the Autonomous Avatar's 1720 a backstory, history and memory.
  • Emotional Attributes 2643 may include the Autonomous Avatar's 1720 a emotional disposition, and reaction and response algorithms.
  • FIG. 27 shows a flow chart illustrating the process for driving the Autonomous Avatar 1720 , shown generally at 2700 .
  • the process begins from step 2702 where the dialog is received by Dialog Manager 1802 .
  • the dialog is received by Dialog Manager 1802 .
  • driving a personal avatar may interact with the Autonomous Avatar 1720 , thereby generating the dialog.
  • any Language Source 103 p may provide the dialog.
  • step 2604 the data source is crawled by the Source Crawler 1804 .
  • the process then proceeds to step 2706 where the received dialog is segmented by utilizing the Segmenter 1712 .
  • step 2608 the dialog is parsed by utilizing the Parser 1713 .
  • step 2710 semantic analysis is performed by utilizing the Semantic Engine 1714 .
  • analytic analysis is performed by utilizing the Analytics Engine 1715 .
  • Pragmatic analysis is then performed by utilizing the Pragmatics Analyzer 1716 , at step 2714 .
  • step 2716 recommendations are generated by utilizing the Recommender 1717 .
  • dialog is generated for the Autonomous Avatar 1720 by utilizing the Response Generator 1718 . The process then ends.
  • FIG. 28 shows a flow chart illustrating the process for dialog segmentation, shown generally at 2706 .
  • the process begins from step 2704 of FIG. 27 .
  • the process then proceeds to step 2802 where particle representation is performed.
  • Particle representation includes organizing the dialog into a data structure by utilizing the Particle Representor 1902 .
  • step 2804 character conversion is performed. Character conversion may utilize the Character Converter 1904 .
  • step 2806 the dialog is split into individual sentences.
  • the Splitter 1906 may be utilized for dialog splitting. The process then concludes by proceeding to step 2708 of FIG. 27 .
  • FIG. 29 shows a flow chart illustrating the process for dialog parsing, shown generally at 2708 .
  • the process begins from step 2706 of FIG. 27 .
  • the process then proceeds to step 2902 where grammar is linked.
  • grammar linking of particular word groups, such as “have” and “being” may be particularly useful for providing precision of the dialog when converted to computer language.
  • Grammatical linking may utilize the Grammar Linker 2002 .
  • step 2904 grammatical mood is determined by utilizing the Grammatical Mood Analyzer 2004 .
  • the process then proceeds to step 2906 where tagging of the dialog is performed.
  • tagging of the dialog provides an abundance of information as to the sentence structure, and ultimately to the meaning of the dialog. Tagging may be performed by the Tagger 2006 .
  • step 2908 variants are pruned by utilizing the Variant Pruner 2008 . Variant pruning may utilize Bayesian statistics or other statistical methodologies. The process then concludes by proceeding to step 2710 of FIG. 27 .
  • FIG. 30 shows a flow chart illustrating the process for semantic analysis, shown generally at 2710 .
  • the process begins from step 2708 of FIG. 27 .
  • the process then proceeds to step 3002 where the dialog is tokenized by utilizing a token frame dictionary lookup.
  • the tokenization may utilize the Token Framer 2102 .
  • step 3004 semantic lexicons are generated by utilizing the Lemmatizer 2104 .
  • a detailed discussion of the generation of semantic lexicons is provided above in the discussion of natural language processing.
  • semantic networking may be performed, by utilizing the Semantic Network Generator 2106 .
  • prototypical semantic networks may be pre-computed for access later by commonalities to the dialog, thereby reducing the processing demands on the Autonomous Avatar Driver 150 .
  • step 3008 co-references are resolved by utilizing the Co-reference Resolver 2108 .
  • Co-references may be resolved in three ways: pattern matching, context appropriateness, and verb reference to a subject.
  • the process then concludes by proceeding to step 2712 of FIG. 27 .
  • FIG. 31 shows a flow chart illustrating the process for analytic analysis, shown generally at 2712 .
  • the process begins from step 2710 of FIG. 27 .
  • the process then proceeds to step 3102 where common words are filtered for by utilizing the Common Word Filter 2202 .
  • Common words may include such regularly used words as “the” and “a”.
  • the process then proceeds to step 3104 where token frame statistics are performed by utilizing the Token Frame Statistics Analyzer 2204 .
  • Token frame statistics may include counting lemmatized words and nodal analysis. However, additional statistical inquiries may be preformed as is desired for Autonomous Avatar Driver 150 functionality.
  • step 3106 the N-gram analysis is performed by utilizing the N-gram Analyzer 2206 .
  • the N-gram analysis includes sliding a window of N words along the dialog to generate statistics on word popularity. The number of N words may vary dependent upon language of the dialog.
  • the process then concludes by proceeding to step 2714 of FIG. 27 .
  • FIG. 32 shows a flow chart illustrating the process for pragmatic analysis, shown generally at 2714 .
  • the process begins from step 2712 of FIG. 27 .
  • the process then proceeds to step 3202 where slang is resolved by utilizing the Slang Resolver 2302 .
  • knowledge may be represented using knowledge templates.
  • the Knowledge Template Resolver 2304 may be utilized to generate the knowledge templates.
  • the process then proceeds to step 3206 where proper nouns may be grouped by utilizing the Proper Noun Grouper 2306 .
  • step 3208 affect may be estimated using a combination of grammatical mood analysis and contextual referencing.
  • Affect estimation may utilize the Affect Estimator 2308 .
  • the process concludes by proceeding to step 2716 of FIG. 27 .
  • FIG. 33 shows a flow chart illustrating the process for recommendation generation, shown generally at 2716 .
  • the process begins from step 2714 of FIG. 27 .
  • the process then proceeds to step 3302 where tag clouds are aggregated by utilizing the Cloud Aggregator 2402 .
  • step 3304 individuals may be clustered into “neighborhoods” by related interest or compatibility. Additional clustering criteria may also be utilized, such as abilities, political ideologies or diversity. Clustering may utilize the Hood Clusteror 2404 .
  • step 3306 individuals may be recommended to one another across a social network by utilizing the Social Network Recommender 2406 .
  • contextual advertising may be recommended by utilizing the Contextual Advertising Generator 2408 .
  • the process then concludes by proceeding to step 2718 of FIG. 27 .
  • FIG. 34 shows a flow chart illustrating the process for avatar dialog generation, shown generally at 2718 .
  • the process begins from step 2716 of FIG. 27 .
  • the process then proceeds to step 3402 where dialog is summarized by utilizing the Summarizer 2502 .
  • This dialog may then be managed at step 3404 by utilizing the Dialog Manager 2504 .
  • the process then ends.
  • the present invention provides an Autonomous Avatar Driver System 100 that provides for advanced dialog management that is highly contextual, and which provides highly relevant recommendation based upon conversation dialog.
  • Such autonomous avatar drivers are useful for driving believable avatars in conjunction with massively multiplayer Online Games, virtual social worlds, online web communities corporate web sites, health care and educational programs and websites.
  • Autonomous avatar driver systems may be entirely software, entirely hardware, or a combination of software and hardware. The advantages of such an efficient system include enhancing the believability of avatars, having knowledge representation from vast data sources, improved searching which utilizes semantics, improved tag clouds, improved corporate relations, advanced recommendation abilities and positive repercussions for the healthcare and education industries.

Abstract

The autonomous avatar driver is useful in association with language sources. A sourcer may receive dialog from the language source. It may also, in some embodiments, receive external data from data sources. A segmentor may convert characters, represent particles and split dialog. A parser may then apply a link grammar, analyze grammatical mood, tag the dialog and prune dialog variants. A semantic engine may lookup token frames, generate semantic lexicons and semantic networks, and resolve ambiguous co-references. An analytics engine may filter common words from dialog, analyze N-grams, count lemmatized words, and analyze nodes. A pragmatics analyzer may resolve slang, generate knowledge templates, group proper nouns and estimate affect of dialog. A recommender may generate tag clouds, cluster the language sources into neighborhoods, recommend social networking to individuals and businesses, and generate contextual advertising. Lastly, a response generator may generate responses for the autonomous avatar using the analyzed dialog. The response generator may also incorporate the generated recommendations.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This is a continuation-in-part of co-pending United States Application Attorney Docket Number HC-0701, application Ser. No. 11/682,813, filed on Mar. 6, 2007, entitled “Systems and Methods for Natural Language Processing”, which is hereby fully incorporated by reference.
  • BACKGROUND OF THE INVENTION
  • The present invention relates to a system and method for autonomous avatar driver, and more particularly an autonomous avatar driver capable of natural language processing for improved data querying and enhanced knowledge representation for avatars, virtual healthcare advisors, virtual personal assistants and non player characters within video games, Massively Multiplayer Online Games (MMOGs), virtual worlds, online social networks, and virtual classrooms. The autonomous avatar driver may also be enabled to mine dialog for data that may then be utilized to generate recommendations to a user.
  • The number of active subscribers to MMOGs is at least 10 million people. Each person pays $15 and up a month to play these games, and maybe and additional 20 million people login occasionally. Estimates are that players spent about $1 billion in real money in 2005 on virtual goods and services for MMOGs combined. Moreover, at least 1.5 million people subscribe to virtual worlds. In January, 2006, inside one such virtual social world, people spent nearly $5 million in some 4.2 million transactions buying or selling clothes, buildings, and the like. Moreover, participants in web communities number in the multiple tens of millions. Additionally, traditional video games have sold over $30 billion in goods since 2000.
  • Many of these games and virtual environments are coming to rely on believable characters as an integral part of the story, and as video game players come to expect increasingly believable scenarios, the characters must improve as well. Integral to believable characters is their ability to comprehend conversation and exhibit a knowledge base.
  • Currently, knowledge representation is fatally limited by resource and financial shortcomings. Data required for knowledge representation may be incorporated into databases in particular formats to be usable for knowledge representation. Such a system is cumbersome, requiring large numbers of man-hours, often with costly specialists, and huge amounts of storage for the data. Infallibly, despite these huge expenditures, the data set created for the knowledge representation will be incomplete, leading to palpable gaps in the “knowledge” of the character.
  • A natural language process that is able to examine external data sources, and glean relevant data may provide far more cost efficient, complete and adaptable data sources for knowledge representation is currently lacking. For games, virtual worlds and narratives, such knowledge representation may become essential to the success of the product.
  • Moreover, personal avatars may likewise benefit from knowledge representation as knowledgeable personal assistants, and more fully developed player characters. The present invention allows for this level of expanded knowledge representation.
  • Additionally, a natural language interface may allow substantially improved searching for relevant frequently asked questions and general data queries. Such improvements have considerable implications for education, healthcare and corporate relations. Current data queries examine matches of words rather than matches of meaning. Natural language processing may improve these searching methods by incorporating semantics.
  • Additionally, such advanced contextual search techniques may be extended to avatar dialog to generate highly targeted data mining. Such data mining may be useful for advertisement, user recommendations and statistical data farming.
  • It is therefore apparent that an urgent need exists for a system and method for an automated avatar driver with contextual recommendation ability. This system would be able to provide highly believable virtual personalities for personal and corporate use.
  • SUMMARY OF THE INVENTION
  • To achieve the foregoing and in accordance with the present invention, systems and methods for an autonomous avatar driver are provided. Such systems and methods are useful for providing a highly believable virtual avatar, or character, with interactive ability and knowledge representation and recommendation ability.
  • The autonomous avatar driver is useful in association with language sources. A sourcer may receive dialog from the language source. It may also, in some embodiments, receive external data from data sources.
  • A segmentor may segment the dialog. The segmentor may include a particle representor, a character converter and a splitter for splitting dialog. A parser may then parse the segmented dialog. The parser may include a linker for linking grammar, a grammatical mood analyzer, a tagger for tagging the dialog and a variant pruner.
  • After parsing the dialog a semantic engine may analyze the parsed dialog for semantics. The semantics engine may include a token framer, a lemmatizer configured to draw from semantic lexicons, a semantic network generator and a co-reference resolver for resolving co-reference ambiguity.
  • An analytics engine may then analyze the semantically analyzed dialog for analytics. The analytics engine may include a common word filter, a token frame statistics engine, and an N-gram analyzer configured to analyze N-grams for word popularity. The token frame statistics engine may count lemmatized words, and perform nodal analysis.
  • Then a pragmatics analyzer may analyze the analytically analyzed dialog for pragmatics. The pragmatics analyzer may include a slang resolver, a knowledge template resolver configured to generate knowledge templates, a proper noun grouper and an affect estimator configured to estimate emotion of the dialog.
  • A recommender may then generate recommendations from the pragmatically analyzed dialog. The recommender may include a cloud aggregator for generating aggregate tag clouds, a hood cluster or for clustering the end users into neighborhoods based upon dialog parsed from them, a social network recommender for recommending social networking to individuals and businesses, and a contextual advertising generator.
  • Lastly, a response generator may generate responses for the autonomous avatar using the pragmatically analyzed dialog. The response generator may also incorporate the generated recommendations from the recommender.
  • Currently “automated” avatars are limited to closed systems of limited referencing data. Moreover, there has been little to no extension of natural language processing into virtual environments for enhanced data retrieval or “intelligent” avatars and virtual personalities. Thus, traditional “automated” avatars are highly limited to narrow contextual situations. These traditional avatars are not believable and limited in functionality. The present invention discloses systems and methods for expanding avatar functionality, utility and believability.
  • Note that the various features of the present invention described above may be practiced alone or in combination. These and other features of the present invention will be described in more detail below in the detailed description of the invention and in conjunction with the following figures.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • In order that the present invention may be more clearly ascertained, one embodiments will now be described, by way of example, with reference to the accompanying drawings, in which:
  • FIG. 1A shows a schematic block diagram illustrating an autonomous avatar driver with a natural language processor system with knowledge representation in accordance with some embodiments of the present invention;
  • FIG. 1B shows an exemplary functional block diagram of the natural language processor in accordance with some embodiments of the present invention;
  • FIG. 2 shows an exemplary functional block diagram of the natural language analyzer of FIG. 1;
  • FIG. 3 shows a flow chart illustrating the natural language processing for the natural language processor system of FIG. 1;
  • FIG. 4 shows a flow chart illustrating the process for processing language for knowledge representation and semantics for the natural language processor system of FIG. 1;
  • FIG. 5 shows a flow chart illustrating the process for preparing language for the natural language processor system of FIG. 1;
  • FIG. 6 shows a flow chart illustrating the process for chunking text for the natural language processor system of FIG. 1;
  • FIG. 7 shows a flow chart illustrating the process for setting grammatical mood for the natural language processor system of FIG. 1;
  • FIG. 8 shows a flow chart illustrating the process for pattern recognition for the natural language processor system of FIG. 1;
  • FIG. 9 shows a flow chart illustrating the process for post operations for the natural language processor system of FIG. 1;
  • FIG. 10 shows a flow chart illustrating the process for preparing processed text for retrieval for the natural language processor system of FIG. 1;
  • FIG. 11 shows a flow chart illustrating the process for an embodiment of data preparation for semantic searches utilizing the natural language processor system of FIG. 1;
  • FIG. 12 shows a flow chart illustrating the process for an embodiment of semantic searches for frequently-asked-questions utilizing the natural language processor system of FIG. 1;
  • FIG. 13A shows a flow chart illustrating the process for one embodiment of determining semantic similarities for semantic searches utilizing the natural language processor system of FIG. 1;
  • FIG. 13B shows a flow chart illustrating the process for another embodiment of determining semantic similarities for semantic searches utilizing the natural language processor system of FIG. 1;
  • FIG. 14 shows a flow chart illustrating the process for an embodiment of a general semantic search utilizing the natural language processor system of FIG. 1;
  • FIG. 15A shows a flow chart illustrating the process for one embodiment of determining semantic similarities for general semantic searches utilizing the natural language processor system of FIG. 1;
  • FIG. 15B shows a flow chart illustrating the process for another embodiment of determining semantic similarities for general semantic searches utilizing the natural language processor system of FIG. 1;
  • FIG. 16 shows a flow chart illustrating the process for generating tag clouds in accordance with an embodiment of the present invention;
  • FIG. 17 shows a schematic block diagram illustrating an autonomous avatar driver system with recommendation ability in accordance with some embodiments of the present invention;
  • FIG. 18 shows a schematic block diagram illustrating a sourcer in accordance with the autonomous avatar driver system of FIG. 17;
  • FIG. 19 shows a schematic block diagram illustrating a segmenter in accordance with the autonomous avatar driver system of FIG. 17;
  • FIG. 20 shows a schematic block diagram illustrating a parser in accordance with the autonomous avatar driver system of FIG. 17;
  • FIG. 21 shows a schematic block diagram illustrating a semantic engine in accordance with the autonomous avatar driver system of FIG. 17;
  • FIG. 22 shows a schematic block diagram illustrating an analytics engine in accordance with the autonomous avatar driver system of FIG. 17;
  • FIG. 23 shows a schematic block diagram illustrating a pragmatics analyzer in accordance with the autonomous avatar driver system of FIG. 17;
  • FIG. 24 shows a schematic block diagram illustrating a recommender in accordance with the autonomous avatar driver system of FIG. 17;
  • FIG. 25 shows a schematic block diagram illustrating a response generator in accordance with the autonomous avatar driver system of FIG. 17;
  • FIG. 26A shows a schematic block diagram of the virtual universe for the autonomous avatar driver system of FIG. 17;
  • FIG. 26B shows a logical block diagram of virtual environments for the autonomous avatar driver system of FIG. 17;
  • FIG. 26C shows a schematic block diagram of a virtual environment for the autonomous avatar driver system of FIG. 17;
  • FIG. 26D shows a schematic block diagram of an autonomous avatar for the autonomous avatar driver system of FIG. 17;
  • FIG. 27 shows a flow chart illustrating the process for driving the autonomous avatar in accordance with some embodiments of the present invention;
  • FIG. 28 shows a flow chart illustrating the process for dialog segmentation in accordance with some embodiments of the present invention;
  • FIG. 29 shows a flow chart illustrating the process for dialog parsing in accordance with some embodiments of the present invention;
  • FIG. 30 shows a flow chart illustrating the process for semantic analysis in accordance with some embodiments of the present invention;
  • FIG. 31 shows a flow chart illustrating the process for analytic analysis in accordance with some embodiments of the present invention;
  • FIG. 32 shows a flow chart illustrating the process for pragmatic analysis in accordance with some embodiments of the present invention;
  • FIG. 33 shows a flow chart illustrating the process for recommendation generation in accordance with some embodiments of the present invention; and
  • FIG. 34 shows a flow chart illustrating the process for avatar dialog generation in accordance with some embodiments of the present invention.
  • DETAILED DESCRIPTION OF THE INVENTION I. Overview of Autonomous Avatar Driver
  • The present invention will now be described in detail with reference to several embodiments thereof as illustrated in the accompanying drawings. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be apparent, however, to one skilled in the art, that the present invention may be practiced without some or all of these specific details. In other instances, well known process steps and/or structures have not been described in detail in order to not unnecessarily obscure the present invention. The features and advantages of the present invention may be better understood with reference to the drawings and discussions that follow.
  • The present invention relates to systems and methods for autonomous avatar driver that provides for enhanced functionality, utility and believability for avatars, and recommendation ability. Such autonomous avatar drivers are useful for generating believable avatars in conjunction with Massively Multiplayer Online Games (MMOGs), virtual social worlds, online web communities, corporate web sites, health care and educational programs and websites. Believability of the avatars is due to their ability to respond to language, also referred to as dialogue, also known in the art as “natural language”, in a spontaneous manner, and their ability to provide an extensive knowledge base, referred to as knowledge representation. Moreover, such natural language processors allow for semantic searches, searches that query not only the words provided but the meaning, thereby making the search more efficient, intuitive and useful.
  • As known to those skilled in the art, an avatar is a virtual representation of an individual within a virtual environment. Avatars often include physical characteristics, statistical attributes, inventories, social relations, emotional representations, and weblogs (blogs) or other recorded historical data. Avatars may be human in appearance, but are not limited to any appearance constraints. Avatars may be personifications of a real world individual, such as a Player Character (PC) within a MMOG, or may be an artificial personality, such as a Non-Player Character (NPC). Additional artificial personality type avatars include personal assistants, guides, educators, answering servers and information providers. Additionally, some avatars may have the ability to be automated some of the time, and controlled by a human at other times. Such Quasi-Player Characters (QPCs) may perform mundane tasks automatically, but more expensive human agents take over in cases of complex problems.
  • The avatar driven by the autonomous avatar driver is generically defined. The avatar may be a character, non-player character, quasi-player character, agent, personal assistant, personality, guide, representation, educator or any additional virtual entity within virtual environments. Avatars may be as complex as a 3D rendered graphical embodiment that includes detailed facial and body expressions, or may be as simple as a faceless, non-graphical widget, capable of limited, or no function beyond the natural language processor. In a society of ever increasing reliance and blending between real life and our virtual lives, the ability to have believable and useful avatars is highly desirable and advantageous.
  • To facilitate discussion, FIG. 1A shows a schematic block diagram illustrating an Autonomous Avatar Driver System 100 with knowledge representation in accordance with some embodiments of the present invention.
  • Central to the Autonomous Avatar Driver System 100 is the Autonomous Avatar Driver 150. The Autonomous Avatar Driver 150 and Natural Language Processor 190 are seen coupled to a Local Area Network (LAN) 106 in a network operations center (NOC). Alternatively, in some embodiments the Natural Language Processor 190 and/or Autonomous Avatar Driver 150 may be located at Language Sources 103 a to 103 m. Moreover, the Autonomous Avatar Driver 150 and/or Natural Language Processor 190 may be located at hosting systems. Host systems may include corporate servers, game systems, MMOG's, websites or other virtual environments. The Language Sources 103 a, 103 b to 103 m may be connected to the LAN 106 through a Wide Area Network (WAN) 101. The most common WAN 101 is the internet; however any WAN, such as a closed gamming network, is intended to be considered as within the scope of the WAN 101. In some embodiments, Language Sources 103 p, 103 q to 103 z may couple directly to the LAN 106. Language Sources 103 a to 103 z provide language to the Natural Language Processor 190 for processing. Language may be in the form of text, or in some embodiments may include audio files and graphical representations, including facial features and body language. The Natural Language Processor 190 includes a Natural Language Analyzer 130, Data Miner 120, Databases 110 and Natural Language Generator 140.
  • A Firewall 104 separates the WAN 101 and Language Sources 103 p, 103 q to 103 z from the LAN 106. The Firewall 104 is coupled to a Server 105, which acts as an intermediary between the Firewall 104 and the LAN 106.
  • Data Sources 102 a, 102 b to 102 x may be coupled to the WAN 101 for access by the Autonomous Avatar Driver 150 and/or Natural Language Processor 190. Data Sources 102 a to 102 x provide external data for use by the Autonomous Avatar Driver 150 and/or Natural Language Processor 190 for the knowledge representation. Additionally, Databases 110 may be coupled to the LAN 106 with for storing frequently utilized data, and for dialogue storage. Together, Data Sources 102 a to 102 x and the Databases 110 provide the statistical information and data required for accurate language processing and knowledge representation. The Databases 110 may include, but are not limited to four child databases. These databases include a dictionary, wordnet, part-of-speech tags (or speech tags), and conversation history. These Databases 110 provide the resources necessary for basic natural language processing with a persistent conversation memory. Additional databases may be included as is desired, since it is the intention of the invention to be amorphic and adaptable as technology and resources become available.
  • The Autonomous Avatar Driver 150 as disclosed may be coupled to the Natural Language Analyzer 130, Data Miner 120 and Natural Language Generator 140. In some embodiments, the Autonomous Avatar Driver 150 may utilize these components to drive the avatar. In some alternate embodiments, the Autonomous Avatar Driver 150 may include alternate components that are capable of language analysis and response.
  • The Natural Language Analyzer 130 may be coupled to a Data Miner 120 and Natural Language Generator 140. The Natural Language Analyzer 130, as will be seen, generates a wealth of data regarding conversations including the meaning, moods and responses to the language, or dialogue. Such data may be extremely valuable for the generation of highly targeted advertising, statistical information, marketing data, sociological research data, political barometer, and many other areas of interest. The Data Miner 120 allows for the searching and congregation of said content-dense data for these purposes.
  • The Natural Language Generator 140 is capable of receiving semantics, in the form of lexical chains, grammatical moods and represented knowledge, to generate an appropriate response. In some embodiments, the Natural Language Generator 140 returns text. In alternate embodiments the Natural Language Generator 140 may return a response in the native form of the Language Source 103 a to 103 z. Thus, in some embodiments voice synthesizer capabilities may be included in the Natural Language Generator 140. In some embodiments, it is advantageous to include a system within the Natural Language Generator 140 that is capable of identifying taboo or undesired language and prevent such language from being part of responses. This feature is very important when dealing with children, or other individuals sensitive to particular conversations. Additionally, the Natural Language Generator 140 may, in some embodiments, be designed to incorporate personalized product placement for advertising purposes when relevant to the conversation. In fact, in some embodiments, the Natural Language Generator 140 may be designed to steer conversations toward product placement, political views, etc. In such a way, avatars incorporating such a Natural Language Processor 190 may become highly specialized and sophisticated advertising media.
  • Additionally, in some embodiments, the Natural Language Generator 140 may include an accompanying graphical generator. Said graphical generator may provide graphical representations, movements, facial features, stance, gestures or other visual stimulus appropriate to the generated language. Such a system may be desirable for very personalized interaction; however, such a system requires large computations and bandwidth to be practical for multiple avatars. As such, said system for graphical generation is best utilized when the Autonomous Avatar Driver System 100 is hosted on an individual's system in order to run a personal avatar, or on a corporate server for the corporate avatar.
  • Alternatively, in some embodiments, the graphical generator may output a set of mood indicators that may then be utilized by the avatar host to generate movements that correspond to the given mood of the language. For example a series of predetermined mood categories may be assigned, and each mood category includes a numerical tag. The host system may have postures, facial features and physical movements corresponding to each of the mood categories. The graphical generator outputs the numerical mood tag, and the host system is then able to produce a range of graphics appropriate to the language. Such a system minimizes computations required by the graphical generator and minimizes required bandwidth. Moreover, in some embodiments, the graphical generator may provide a series of graphical indicators to more finely tune graphical representations. For instance the graphical generator may output a first variable relating to the general mood, and subsequent variables, each more finely tuning the graphical representations. The advantage of such a system is that the complexity of host graphical representations may vary greatly in their level of detail and range of emotional states. As such, the host system may utilize outputs, and ignore subsequent tuning variables as they become ineffectual within said host system. The advantage of such a system is the ability of a standard output by the graphical generator, yet having each host system receiving the level of graphical detail desired.
  • II. Natural Language Processor System and Method
  • FIG. 1B shows an exemplary functional block diagram of the Natural Language Processor 190 as involved in the Autonomous Avatar Driver System 100 of FIG. 1. Exemplary Language Source 130 a may be seen providing language to the Natural Language Processor 190. The language is received by the Natural Language Analyzer 130, which analyzes the language for semantics. The Natural Language Analyzer 130 is coupled to the Natural Language Generator 140, the Databases 110 and the Data Miner 120. Moreover, the exemplary Data Source 102 a provides data to the Natural Language Analyzer 130. The Natural Language Analyzer 130 may utilize the Databases 110 for analyzing the language semantics. Additionally, the Natural Language Analyzer 130 may provide semantic information from the analyzed language to the Data Miner 120 and the Natural Language Generator 140.
  • The Natural Language Generator 140 is coupled to the Natural Language Analyzer 130, the Databases 110 and the Data Miner 120. Moreover, the exemplary Data Source 102 a provides data to the Natural Language Generator 140. The Natural Language Generator 140 may utilize the Databases 110 for generating return language. Additionally, the Natural Language Generator 140 may provide semantic information from the generated language to the Data Miner 120. The generated language may be provided to the exemplary Language Source 130 a. In this way, the Natural Language Processor 190 is capable of semantically relevant conversation, with knowledge representation.
  • The Data Miner 120, aside from gaining semantic information from the Natural Language Generator 140 and Natural Language Analyzer 130 may query the Databases 110 in some embodiments. The Data Miner 120 is capable of providing the mined data to Data Consumers 125. Such Data Consumers 125 include research groups, advertisers, educators, marketing analysts, political analysts, quality control analysts, and opinion based review groups, among many others. Such Data Consumers 125 may provide an important source of revenue in some embodiments, as well as providing service data for the end users.
  • FIG. 2 shows a functional block diagram of the Natural Language Analyzer 130 of FIG. 1. Within natural language processing there are terms well known to those skilled in the art that are worth a cursory definition. A ‘lemma’ is a set of lexical forms with the same stem, the same part-of-speech and the same word-sense. A ‘lexeme’ is a pairing of a particular orthographic form with a symbolic meaning representation. Moreover, a finite list of lexemes is referred to as a ‘lexicon’. Lastly, ‘types’ refers to the total number of distinct words in a corpus.
  • A Receiver 201 receives the Language from the Language Source 103 a to 103 z from over the network. The Receiver 201 may also identify the format of the language. In some embodiments the Receiver 201 may include a spelling checker; however, as it will be seen, improper grammar, slang and typos are anticipated and dealt with by the Natural Language Analyzer 130. The Receiver 201 is coupled to the Language Preparer 220 which includes the Language Converter 221, Sentence Detector 222, Part-of-Speech Tagger (or Speech Tagger) 223 and Speech Parser (or Parser) 224. The Receiver 201 couples directly with the Language Converter 221 and the Sentence Detector 222. When the received dialogue, or language, is in a native format compatible with the Natural Language Analyzer 130 the Receiver 201 may provide the language directly to the Sentence Detector 222. In some embodiments the native language for the Natural Language Analyzer 130 includes text files. Alternatively, if the corpus is in a nonnative format, such as a rich text, audio file or graphical file, the dialog is first analyzed by the Language Converter 221. In some embodiments, the Language Converter 221 may receive a plurality of language formats and convert them to a text format. In said Language Converter 221, voice recognition software may be required. Additionally, in some embodiments the Language Converter 221 may include image recognition in order to interpret body language, sign language and facial features for conversion into the native language for the Natural Language Analyzer 130. After nonnative language is analyzed by the Language Converter 221 the text equivalent of the language is processed by the Sentence Detector 222.
  • It should be mentioned that the instant invention is designed to be able to analyze, mine and generate in a variety of languages. The base architecture of the Autonomous Avatar Driver System 100 is nearly identical regardless of whether the Language Sources 103 a to 103 z are in English, German or Japanese, to name a few applicable languages. The instant invention utilizes statistical learning, dictionaries and grammatical rules; as such, as long as the appropriate training corpuses are available the system may be utilized across a variety of languages.
  • In some embodiments a translation matrix may be utilized in order to provide semantic consistency, and when it is desirous for the input language format to differ from the output language format. Such a translation matrix may exist on the front end, translating the incoming language. Alternatively, in some embodiments, said translation matrix may exist after the Natural Language Analyzer 130 in order to translate the semantic meaning of the received language. Alternatively, in some embodiments, said translation matrix may exist before the Natural Language Generator 140, thereby enabling all analysis to be performed within the native language format, and only language generation is performed in the output language format. Alternatively, in some embodiments, said translation matrix may exist after the Natural Language Generator 140, thereby enabling all natural language processing to be performed within the native language format, and only language output is in the output language format. Moreover, a combination of translation matrices may be utilized in some embodiments.
  • The Sentence Detector 222 is coupled with the Language Converter 221 and the Part-of-Speech Tagger (or Speech Tagger) 223. The Sentence Detector 222 utilizes algorithms to determine separate sentences. The sentences are made up of elements, which include words and symbols. The Speech Tagger 223 first cleans up the raw text by tokenizing it. Tokenization includes replacing certain elements with tokens. The tokens delineate meanings of the elements that they replace. Additionally, tokenization ensures that special characters are separated from the words in the sentence so that the elements may be parsed easier.
  • The Speech Tagger 223 then annotates each element with a tag that classifies the element with pre-defined categories. The Speech Tagger 223 is usually trained with a training set of data since it relies upon statistics in order to determine the meaning of ambiguous words. There are different standards for Part-of-Speech tags; however, in some embodiments the Penn Treebank structure is utilized. Table 1 shows the tags of the Penn Treebank.
  • TABLE 1
    Part-of-Speech Tags
    # Tag Part-of-Speech
    1. CC Coordinating conjunction
    2. CD Cardinal number
    3. DT Determiner
    4. EX Existential there
    5. FW Foreign word
    6. IN Preposition, subordinating
    conjunction
    7. JJ Adjective
    8. JJR Adjective, comparative
    9. JJS Adjective, superlative
    10. LS List item marker
    11. MD Modal
    12. NN Noun, singular or mass
    13. NNS Noun, plural
    14. NP Proper noun, singular
    15. NPS Proper noun, plural
    16. PDT Predeterminer
    17. POS Possessive ending
    18. PP Personal pronoun
    19 PP$ Possessive pronoun
    20. RB Adverb
    21. RBR Adverb, comparative
    22. RBS Adverb, superlative
    23. RP Particle
    24. SYM Symbol
    25. TO to
    26. UH Interjection
    27. VB Verb, base form
    28. VBD Verb, past tense
    29. VBG Verb, gerund or present
    participle
    30. VBN Verb, past participle
    31. VBP Verb, non-3rd person singular
    present
    32. VBZ Verb, 3rd person singular
    present
    33. WDT Wh-determiner
    34. WP Wh-pronoun
    35. WP$ Possessive wh-pronoun
    36. WRB Wh-adverb
  • The Speech Tagger 223 may output the tagged sentence via the Outputter 203, or may continue through the Speech Parser 224, which is coupled to the Speech Tagger 223. Speech Parser 224 chunks, or parses, the sentences into discrete, non-overlapping grammatical chunks. Each chunk is a partial structure of the sentence. Chunking is important for downstream semantic analysis and pattern recognition. After Chunking is performed, the chunks are annotated for property based searching and improved information extraction. Examples of chunking rules utilized by the Speech Parser 224 are illustrated below in Table 2:
  • TABLE 2
    Chunk Types by Annotation
    Annotation Chunk Type Definition
    S Sentence
    NP Noun Phrase
    VP Verb Phrase
    PP Prepositional Phrase
    AP Adjective Phrase
    S NP + VP
    VP Verb
    VP Verb + NP
    VP Verb + NP + PP
    VP Verb + PP
    PP Preposition + NP
    NP Pronoun/Proper Noun/Det Nominal
    NP (Det) (Card) (Ord) (Quant) (AP) Nominal
    Nominal Noun Nominal/Noun/Nominal PP (PP) (PP)
    Nominal Nominal GerundVP
    GerundVP GerundV + NP/GerundV + PP/Gerund + Verb/Gerund +
    Verb + NP + PP
  • The Speech Parser 224 may output the chunked sentence via the Outputter 203 or may continue through the Semantic Analyzer 230, which is coupled to the Speech Parser 224. The Semantic Analyzer 230 includes a Grammar Analyzer 231, a Referencer 232, an Affect Analyzer 233 and a Pattern Recognizer 234. The Grammar Analyzer 231 is coupled to the Speech Parser 224, the Referencer 232 and the Outputter 203. Grammar Analyzer 231 determines the grammatical mood of the sentence. Examples of grammatical moods assigned by the Grammar Analyzer 231 are illustrated below in Table 3:
  • TABLE 3
    Grammatical Moods
    Grammatical Mood Definition
    Declarative structure NP + VP
    Imperative structure VP
    Interrogative structure (yes/no) Aux NP + VP
    Interrogative structure Wh-NP + VP/Wh-NP + Aux NP + VP
    (who/what/where/when/why)
  • In some embodiments, additional definitions of grammatical moods may be utilized by the Grammar Analyzer 231; however in some other embodiments the grammatical moods illustrated in Table 3 are sufficient, and any dialogue that does not fit into one of the previous definitions may be assumed to be a declarative grammatical mood. Such an assumption generally proves computationally effective, as well as accurate enough for most circumstances.
  • Of course, in some embodiments the Grammar Analyzer 231 may utilize a statistical engine, which may incorporate training, in order to statistically determine the grammatical mood, as apposed to a rule based engine as described above. Alternatively, a hybrid rule and statistical engine may be utilized, wherein the rules are utilized for simple sentence structures, but learned statistical methods may be utilized for more complicated or less determinable sentences. Such a system may be of particular use when slang is highly prevalent within the language. The advantages of such a system include the incorporation of learning, or statistical tuning, in order to improve accuracy.
  • The Grammar Analyzer 231 is also coupled to the Outputter 203, such as to output the grammatical mood of the sentence. Grammatical moods may result in prioritization of discourse; for example imperative structured sentences may require an action or trigger an event.
  • The Referencer 232 may reference dictionaries and thesauruses in order to reference the elements of the sentence for semantics. Additionally, in some embodiments, the Referencer 232 may cross reference graphical cues, such as gestures, with language pronouns in order to delineate the meaning of ambiguous pronouns. The Referencer 232 is then coupled to the Affect Analyzer 233.
  • The Affect Analyzer 233 utilizes the results of the Referencer 232 and the grammatical mood from the Grammar Analyzer 231 in order to infer the affect of the sentence. Graphical recognition may be of particular use by the Affect Analyzer 233 to infer the language affect. Additionally, in some embodiments, tactic responses may be utilized to analyze affect. Tactic responses, as is known to those skilled in the art, includes the usage of information, such as the rate of typing, change in voice pitch, thermal readings of skin, electrical conductivity of skin, etc. in order to help infer the users' emotional state for the purposes of affect analysis. Such tactic response devices may be as simple as analyzing the rhythm of typing, or may include external peripheral devices worn by the users. The Affect Analyzer 233 may provide cues as to the appropriate disposition of return language to the Natural Language Generator 140. Results from the Affect Analyzer 233 may proceed directly to the Knowledge Representor 202 or may proceed through the Pattern Recognizer 234.
  • Pattern Recognizer 234 utilizes statistics, through a statistical engine, to determine meaning of the language. As such, the Pattern Recognizer 234 may be trained. In some embodiments it may be advantageous to train the Pattern Recognizer 234 within the environment that it will function, in order to incorporate the slang of that environment. Within the gamming community entire dialects of speech have developed, often differing between game to game, and even separate factions within single games. It would be advantageous to have the Natural Language Analyzer 130 enabled to process slang, and the Natural Language Generator 140 configured to use said slang, in order to enhance the believability of the avatar. Additionally, it is important to note that these dialects and slang language adapt rapidly, therefore a constant training of the Pattern Recognizer 234 may be beneficial. The Pattern Recognizer 234 may generate a lexical chain in order to convey the meaning of the language dialogue. Lexical chains, well known to those skilled in the art, are strings of words designed to convey meanings of bodies of text, and are designed to eliminate incongruities and ambiguities. The lexical chains generated by the Pattern Recognizer 234 may be processed through the Knowledge Representor 202.
  • The Knowledge Representor 202 utilizes knowledge templates in order to query Data Sources 102 a to 102 x and Databases 110 in order to provide a more enriching and believable conversation with the avatar. The Databases 110 may include previous conversation logs, providing a level of personal ‘memory’ for the avatar. In this way, previous conversations and happenings may be ‘recalled’ by the avatar. Moreover, external Data Sources 102 a to 102 x may provide a vast wealth of information about virtually any topic, upon which the avatar may draw from in order to effectuate a believable dialogue. Data Sources 102 a to 102 x may include reference sources such a Wikipedia® and IMDb®. Due to the relatively standardized formatting of such Data Sources 102 a to 102 x, knowledge templates may be uniquely designed to access a particular Data Sources 102 a to 102 x. However, once the knowledge template for any Data Sources 102 a to 102 x is generated, any topic contained within the Data Sources 102 a to 102 x may be analyzed by the knowledge template to provide meaningful information for the Knowledge Representor 202 of the Natural Language Analyzer 130. The results of the Knowledge Representor 202 may be sent to the Outputter 203 for downstream applications.
  • FIG. 3 shows a flow chart illustrating the natural language processing shown generally at 300. In step 310 the language received from the Language Source 103 a to 103 z is analyzed for knowledge representation and semantics by the Natural Language Analyzer 130. Then, in step 315, a decision is made whether to perform data mining on the analyzed language. If data mining is desired, then the semantics of the analyzed language are mined for the pertinent data, at step 320. Such data may then be stored for future uses or supplied to the Data Consumers 125. The process then proceeds to step 325, where a decision is made whether to store the results of the natural language analysis.
  • Otherwise, if at step 315, data mining is not desired, the process proceeds directly to step 325, where a decision is made whether to store the results of the natural language analysis. If it is decided to store the natural language analysis at step 325, then the semantics may be stored, or archived, within the Databases 110 at step 330. Stored analyzed language may be subsequently data mined. Also, in some embodiments, the stored analyzed language may be retrieved in future interactions, providing an avatar with a ‘memory’ of previous conversations, and thus resulting in enhanced believability of the avatar. Also, the stored analyzed language may be subsequently cross referenced in order to provide semantics for future ambiguities. For example, the use of ambiguous pronouns in future language may rely upon the previous language semantics in order to resolve said ambiguity. Often these stored semantics may be annotated with the time, location, identity of Language Source 103 a to 103 z and any additional relevant information. In some embodiments, the entire corpus of the received language may be linked to the semantic analysis results. In such a way, a subsequent search may be performed on semantics, and the relevant conversations may be received. The process then proceeds to step 335, where a decision is made whether to generate semantically related return language.
  • Otherwise, if at step 325, storing the results of the natural language analysis is not desired, the process proceeds directly to step 335, where a decision is made whether to generate semantically related return language. If generating semantically related return language is desired, the Natural Language Generator 140 may generate language, at step 340, that utilizes the semantics and knowledge representation from the Natural Language Analyzer 130, from step 310, along with cross referenced information from the Databases 110. Said generated semantically related language may then be output to the Language Source 103 a to 103 z. In this manner return dialogue may be provided, enabling believable conversation.
  • The process then proceeds to step 345, where a decision is made whether to store the generated semantically related language. If storing the generated language is desired, the generated language may be stored, or archived, in the Databases 110 at step 350. Similarly to the analyzed language, stored generated language may be utilized for subsequent data mining and cross referencing. After storage of the generated language, the process ends.
  • Else, if at step 345, storing the generated language is not required, then the process ends.
  • Else, if at step 335, generating semantically related language is not desired, the process ends.
  • FIG. 4 shows a flow chart illustrating the process for analyzing language for knowledge representation and semantics shown generally at 310. In step 410 the language is received. Then, at step 420, the language is prepared. Preparation includes conversion, sentence identification and tokenization. At step 425 a decision is made whether the received language will be understood for semantics, or simply analyzed for knowledge representation.
  • If semantic understanding is desired, the language is tagged at step 430. In some embodiment, Part-of-Speech tagging occurs according to Table 1. The part-of-speech tags database is utilized for tagging. At step 440 the text is chunked, or parsed, into the non-overlapping substructures of the sentence. At step 450, each sentence in the language is analyzed for grammatical mood. At step 460, the text is referenced across dictionaries, WordNet, and any applicable reference corpus. Then, at step 470, the affect of the text is analyzed. Semantics are determined at step 480 by utilizing statistical pattern recognition. Then knowledge representation is performed at step 490. In FIG. 4, additional language analyses may occur after step 490, and before outputting in step 499. The illustration is intended to reflect this adaptability of language analysis. The result of the natural language analysis is then output at step 499.
  • Else, if semantic understanding is not required at step 425, then knowledge representation is performed at step 490. In FIG. 4, additional language analyses may occur after step 490, and before outputting in step 499. The illustration is intended to reflect this adaptability of language analysis. The result of the natural language analysis is then output at step 499.
  • At this point, it should be noted that the Natural Language Processor 190, and methods thereof, disclosed within this application are very modular in nature. Where some embodiments including all of the stated components and steps are discussed, additional embodiments may exist that do not include all of the disclosed components and method steps. Similarly, the components and steps for the Natural Language Processor 190 listed are not intended to be an exhaustive list, as additional refinements become feasible and necessary.
  • FIG. 5 shows a flow chart illustrating the process for preparing language shown generally at 420. The beginning of this process is from step 410, of FIG. 4. At step 521, an inquiry is made as to if the language is in a text format. If the language is in an acceptable text format, the text sentences are separated, at step 523. Then, at step 524, the text is tokenized, wherein any special characters are removed and replaced with meaningful tokens. Tokenization cleans the text for downstream processing. Then the process ends by returning to step 430, of FIG. 4.
  • Else, if the language is not in an acceptable text format, at step 521, the language is converted into an acceptable text format at step 522. Text-to-text format conversions are relatively rapid and straightforward; however, it is intended in some embodiments of the invention, to also receive audio or waveform files and utilize speech recognition software in order to convert the language into text. Moreover, in some embodiments of the invention graphical input is intended to be received, such as sign language, body language and facial expressions. As such, a image recognition system may be utilized to convert graphical information into text for semantic analysis. The text sentences are then separated, at step 523. Then, at step 524, the text is tokenized, wherein any special characters are removed and replaced with meaningful tokens. Then the process ends by returning to step 430, of FIG. 4.
  • FIG. 6 shows a flow chart illustrating the process for chunking text shown generally at 440. The beginning of this process is from step 430, of FIG. 4. At step 611, a placeholder N is set to a base value of 1. Then at step 612 the placeholder N is compared to the number of sentences in the text. If N is greater that the total sentences in the text, then the process ends by returning to step 450 of FIG. 4. This ensures that all sentences within the corpus are chunked prior to the progression of the natural language processing.
  • Otherwise, if, at step 612, N is less than the total number of sentences within the text, a second placeholder M is set to 1 at step 613. Then at step 614 the placeholder M is compared to the number of phrases in sentence N. If M is greater than the total phrases in the sentence N, then 1 is added to N in step 615. This ensures that all phrases within sentence N are chunked prior to the chunking of subsequent sentences. The process then continues at step 612, where the increased placeholder N is compared to the number of sentences in the text.
  • Else, if, at step 614, M is less than the total phrases in sentence N, a query is made to determine if phrase M is a sentence at step 621. The rules utilized to chunk may be referred to above at Table 2. If the phrase M is a sentence, it is annotated as such at step 631. Then at step 616, 1 is added to M and the process then continues to step 614, where the increased placeholder M is compared to the number of phrases in sentence N.
  • Otherwise, if, at step 621 the phrase is not a sentence, a query is made to determine if phrase M is a verb phrase at step 622. If the phrase M is a verb phrase, it is annotated as such at step 632. Then at step 616, 1 is added to M and the process then continues to step 614, where the increased placeholder M is compared to the number of phrases in sentence N.
  • Otherwise, if, at step 622 the phrase is not a verb phrase, a query is made to determine if phrase M is a noun phrase at step 623. If the phrase M is a noun phrase, it is annotated as such at step 633. Then at step 616, 1 is added to M and the process then continues to step 614, where the increased placeholder M is compared to the number of phrases in sentence N.
  • Otherwise, if, at step 623 the phrase is not a noun phrase, a query is made to determine if phrase M is a prepositional phrase at step 624. If the phrase M is a prepositional phrase, it is annotated as such at step 634. Then at step 616, 1 is added to M and the process then continues to step 614, where the increased placeholder M is compared to the number of phrases in sentence N.
  • Otherwise, if, at step 624 the phrase is not a prepositional phrase, a query is made to determine if phrase M is an adjective phrase at step 625. If the phrase M is an adjective phrase, it is annotated as such at step 635. Then at step 616, 1 is added to M and the process then continues to step 614, where the increased placeholder M is compared to the number of phrases in sentence N.
  • Otherwise, if, at step 625 the phrase is not an adjective phrase, a query is made to determine if phrase M is a nominal phrase at step 626. If the phrase M is a nominal phrase, it is annotated as such at step 636. Then at step 616, 1 is added to M and the process then continues to step 614, where the increased placeholder M is compared to the number of phrases in sentence N.
  • Otherwise, if, at step 626 the phrase is not a nominal phrase, a query is made to determine if phrase M is a gerund verb phrase at step 627. If the phrase M is a gerund verb phrase, it is annotated as such at step 637. Then at step 616, 1 is added to M and the process then continues to step 614, where the increased placeholder M is compared to the number of phrases in sentence N.
  • Else, if, at step 627 the phrase is not a gerund verb phrase, 1 is added to M at step 616. The process then continues to step 614, where the increased placeholder M is compared to the number of phrases in sentence N.
  • FIG. 7 shows a flow chart illustrating the process for setting grammatical mood shown generally at 450. The beginning of this process is from step 440, of FIG. 4. At step 751, a placeholder P is set to a base value of 1. Then at step 752 the placeholder P is compared to the number of sentences in the text. If P is greater that the total sentences in the text, then the process ends by returning to step 460 of FIG. 4. This ensures that all sentences within the text are analyzed for grammatical mood prior to the progression of the natural language processing.
  • Otherwise, if, at step 752, P is less than the total number of sentences within the text, the grammatical mood of sentence P is analyzed at step 753. At step 754, a query is made whether sentence P is an interrogative sentence. The definition for an interrogative sentence may be found at Table 3 above. Alternatively, a rule, statistical or hybrid identification system may be utilized as is discussed above. If the sentence is an interrogative sentence, the grammatical mood of sentence P is set to an interrogative structure at step 755. Then, at step 759, 1 is added to P and the process then continues to step 752, where the increased placeholder P is compared to the total number of sentences within the text.
  • Otherwise, if, at step 754, sentence P is not in an interrogative structure, a query is made whether sentence P is an imperative sentence at step 756. The definition for an imperative sentence may be found at Table 3 above. Alternatively, a rule, statistical or hybrid identification system may be utilized as is discussed above. If the sentence is an imperative sentence, the grammatical mood of sentence P is set to an imperative structure at step 757. Then, at step 759, 1 is added to P and the process then continues to step 752, where the increased placeholder P is compared to the total number of sentences within the text.
  • In FIG. 7, additional grammatical mood inquiries may occur after step 756, and before setting the sentence mood to declarative in step 758. Additionally, other mood types may be incorporated within some embodiments. The illustration is intended to reflect this adaptability of grammatical mood analysis.
  • Else, if, at step 756, sentence P is not in an imperative structure, the grammatical mood of sentence P is set to a declarative structure at step 758. This provides that the declarative structure is default for both true declarative sentences, and in the case of sentences whose grammatical mood is difficult to ascertain. Then, at step 759, 1 is added to P and the process then continues to step 752, where the increased placeholder P is compared to the total number of sentences within the text.
  • FIG. 8 shows a flow chart illustrating the process for pattern recognition shown generally at 480. The beginning of this process is from step 470, of FIG. 4. At step 881 synonyms are identified for the elements of the text. Synonyms are different lexemes with the same meaning. Often, words are not synonyms for each other in any context but they are context bound. Then, at step 882, homonyms are identified for the elements of the text. Homonyms are words of the same orthographic and/or phonographic form with unrelated meaning e.g. mouse (computer input device) and mouse (animal). This type of relationship is typically, from a semantic perspective, of little interest. At step 883 hyponyms are identified for the elements of the text. Hyponyms are the more specific lexeme of a pair of lexemes, when one lexeme denotes a subclass of the other. For example, a ‘human’ is a hyponym of ‘primate’. At step 884 hypernyms are identified for the elements of the text. Hypernyms are the less specific lexeme of a pair of lexemes, when one lexeme denotes a subclass of the other. Following the previous example, a ‘primate’ is a hypernym of ‘human’. Synonyms, hyponyms, hypernyms and homonyms are identified through the referencing of the elements through dictionaries and thesauruses as well as statistical pattern recognition.
  • In FIG. 8, additional semantic analyses may occur after step 884, and before lexical chain generation in step 889. The illustration is intended to reflect this adaptability of semantic analysis.
  • At step 889 lexical chains are generated for the text. As stated earlier, lexical chains are data-structures that represent the semantics of a given text at a very abstract level. A lexical chain is made up of a set of different nouns and/or verbs that represent the semantics of a given text. They occur in any order. Moreover, a text may usually be represented by more than one lexical chain. After lexical chains are generated, the process ends by proceeding to step 490 of FIG. 4.
  • FIG. 9 shows a flow chart illustrating the process for outputting post operations shown generally at 499. The beginning of this process is from step 490, of FIG. 4. At step 901 a query is made whether data mining is to be performed upon the semantically analyzed language. If data mining is desired, a data report is generated, at step 990, wherein trends, semantics, ad targets and product placement suggestions are compiled for Data Miner 120 consumption. The generated data report may be personalized to the Language Source 103 a to 103 z providing the dialogue, or language. The semantic analysis data report is output to the Data Miner 120, in step 991. Then the process proceeds to step 902.
  • Else if data mining is not desired at step 901, the process progresses to step 902, where a query is made as to whether a response is required. Grammatical mood plays an important role in the determination of when responses are required. If a response is required, the semantic analysis is output to the Natural Language Generator 140 in step 992. Then the process proceeds to step 903.
  • Else if a response is not desired at step 902, the process progresses to step 903, where a query is made as to whether persistent knowledge of the conversation is required. The type of avatar, content of conversation and length of conversation may all play a role in the persistence of the dialogue. For example, in a game scenario it may be advantageous for an avatar to “remember” lengthy or involved conversation, but “forget” smaller and insignificant dialogue. Not only does this provide the avatar with a believable behavior scheme, but less data-rich content (for data mining purposes) may be eliminated, thus reducing the resources required for storage. If persistent knowledge is required, the semantic analysis is formatted for retrieval in step 993. Then, in step 994, the formatted semantic analysis is stored within the conversation history database. The process then ends. Else if persistent knowledge is not desired, at step 903, the process ends.
  • FIG. 10 shows a flow chart illustrating the process for formatting processed text for retrieval shown generally at 993. The beginning of this process is from step 903, of FIG. 9. At step 1001 the processed sentences are separated. Then, at step 1002, the sentences are chunked by a hash table, wherein the hash table includes the Part-of-Speech tags and words, thus allowing for rapid retrieval of the processed conversation. Additionally, stored conversation may be referenced by the conversant. After hash table formatting, the process ends by proceeding to step 994 of FIG. 9.
  • III. Specific Implementations of Natural Language Processor
  • The following FIGS. 11 to 16 disclose specific implementations of the Natural Language Processor 190 discussed above in FIGS. 1 to 10. While the methods disclosed in FIGS. 11 to 16 illustrate specific implementations of some embodiments, these figures and descriptions are intended to be exemplary in nature.
  • FIG. 11 shows a flow chart illustrating the process for an embodiment of data preparation for semantic searches, shown generally at 1100, utilizing the Natural Language Processor 190 of the Autonomous Avatar Driver System 100. The illustrated method in FIG. 11 is intended to be useful in conjunction with a Frequently Asked Question (FAQ) style data set. A FAQ typically includes a set of predefined questions, followed by an answer to said question. As such, this embodiment of the semantic searching has particularly important implications for corporate customers and guide avatars. For example, a corporation could, in some embodiments, replace its existing FAQ page on its website with a helper avatar. The avatar may be graphically represented, but may often have no physical embodiment. In such embodiments, there may be a field where the inquirer, the user, may type her question. Such an embodiment may mimic a chat style conversation that currently occurs with a customer representative. The avatar may then perform the semantic search of the FAQ database to return the question-answer pairs that are the most relevant to the inquiry. Current help menus provide searching for troubleshooting purposes, however these searches are greatly limited in their ability to comprehend semantics, and inability to process natural language dialogue. When the inquirer is technically sophisticated, these hurdles may be minor due to the proficient usage of search terms. However, many corporations cater to less technically sophisticated clientele, and the ability for these companies to outreach effectively to their customers may be greatly enhanced by allowing natural language searches that often provide better search results than a traditional troubleshooting search; even one performed by a sophisticated and proficient searcher. Moreover, within the gamming and entertainment environments, guide characters, or tutorial characters, may be desired. In these circumstances, it is of particular importance to be able to interpret natural language and provide meaningful responses for the avatar to be believable, and to enhance the experience for the user.
  • The process begins at step 1101, where the answer data is queried. The answer data is the answer portion of a FAQ question-answer pair. Each answer is queried, and in step 523 sentence detection is performed upon the answer data. Then, in step 524, the answer data is tokenized to clean up the answer data by removing all special character and replacing them with meaningful tokens. The parts of speech are tagged in step 430, utilizing, in some embodiments, the tags defined in Table 1. The answer data is then chunked, in step 440, in a manner similar to that previously discussed, and in reference to Table 2. After chunking the process diverges from the prototypical method of natural language processing disclosed earlier in the present invention. At step 1106, the nouns and verbs of the answer data are extracted. Then the question portion of the FAQ question-answer pair is inputted at step 1107. The semantic similarity between the noun and verbs extracted from the answer, and the question is then determined in step 1108.
  • Semantics of the answers is performed by determining the semantics of the nouns and verbs extracted through cross referencing the words and through pattern recognition. Similarly, the semantics of the inputted question are determined through cross referencing the words and pattern recognition. The overlap between the semantic analysis of the question, and semantic analysis of the noun-verb list of the answer, results in the semantic similarity.
  • The semantic similarity is then utilized in 1109 to generate a lexical chain. Then, in step 1110, the generated lexical chain is stored in reference to the question-answer pair it originated from. This process is, in some embodiments, performed for all question-answer pairs. Moreover, this procedure is only required to be completed once, and repeated only whenever new data is introduced in the FAQ. It should be noted that there are many methods of preparing a lexical chain for a question-answer pair in a FAQ; however, due to the semantically similar relationship between the question and its answer the abovementioned process develops a finely tuned lexical chain that is highly relevant to the core meanings of the FAQ pair. By storing these lexical chains, referenced back to their original question-answer pairs, subsequent inquiries may leverage these chains to provide meaningful semantic searches, as may be seen in FIG. 12.
  • FIG. 12 shows a flow chart illustrating the process for an embodiment of semantic searches for frequently-asked-questions, shown generally at 1200. This process follows the process illustrated generally at 1100 at FIG. 11.
  • The instant process begins at step 1201 when an interactive question is inputted. Interactive question refers to a question originating from an end-user of the avatar. This end-user may be the corporation's customer accessing their FAQ on the website, or may be a gamer interacting with a guide avatar. These interactive questions do not, in these embodiments, refer to the question portions of the FAQ.
  • Then, at step 310, the interactive question is processed for semantics and knowledge representation by the Natural Language Analyzer 130 through a process similar to that disclosed at FIG. 4. The results of the semantic analysis of the question are then compared to the lexical chains generated in step 1109 of FIG. 11 for semantic similarity at step 1203. The lexical chains with the greatest semantic similarity to the semantics of the interactive question may then be referenced to the question-answer pair that gave birth to it. These question-answer pairs may be displayed to the end user. Alternatively, only the answer portion of the question-answer pairs may be displayed to the end user. Moreover, in some embodiments the answer may be displayed after being reprocessed through the Natural Language Analyzer 130 and the Natural Language Generator 140 in order to provide an answer that is uniquely tailored to the interactive question, and that may include knowledge representation. In such a way, a guide character, traditionally avoided by gamers due to its daunting monotony and longwinded rigidity, may provide a continuation of the game narrative that is believable and enjoyable to the gamer.
  • FIG. 13A shows a flow chart illustrating the process for one embodiment of determining semantic similarities for semantic searches shown generally at 1203A. The beginning of this process is intended to be from step 1202, of FIG. 12. However, this process 1203A is not limited to such context, and may be employed in many suitable semantic similarity comparisons. At step 1301 a placeholder T is set to a base value of 1. Then at step 1302 the placeholder T is compared to the number of lexical chains that were generated in step 1109 of FIG. 11 for the question-answer pairs of the FAQ. If T is greater that the total number of lexical chains, then the process proceeds to step 1311 where lexical chains are ranked by total weights. This ensures that all lexical chains generated for the FAQ are compared prior to the ranking of semantic similarities.
  • Otherwise, if, at step 1302, T is less than the total number lexical chains of the FAQ, a second placeholder U is set to 1 at step 1303. Then at step 1304 the placeholder U is compared to the number of elements in lexical chain T. If U is greater than the total elements in the lexical chain T, then the weight of the lexical chain T is multiplied by a modifier ZT at step 1308. Modifier ZT provides for corrections to be made for differences between lexical chains, such as for chain length. For example, a highly refined question-answer pair will tend to produce a relatively short lexical chain, since the subject matter explored by the FAQ entry is refined and specific. However, a broad question and answer pair may result in a very long lexical chain. In some circumstances, the interactive question may have more aggregate similarities with the long lexical chain due to its broad nature; however the interactive question is better answered by the narrower, in-depth FAQ entry. In such an instance the modifier ZT may be applied to increase the weight of the shorter lexical chain's semantic similarities.
  • Then a value of 1 is added to T in step 1309. This ensures that all elements within the lexical chain T are compared to the semantics of the interactive question. The process then returns to step 1302, where the increased placeholder T is compared to the number of lexical chains for the FAQ.
  • Else, if, at step 1304, U is less than the total elements in the lexical chain T, then, at step 1305, a weight is assigned to element U by comparing the semantics of the interactive question to element U. A direct semantic match may generate a higher weight value to element U than a more attenuated semantic relationship. At step 1306 the weight assigned to element U is added to the total weight of lexical chain T. Then, at step 1307, a value of 1 is added to U. The process then returns to step 1304, where the increased placeholder U is compared to the number of elements of the lexical chain T. In this manner the total weight of the lexical chain T is the sum of the weights of its respective elements.
  • After all lexical chains have been processed for their total weights; the lexical chains may be ranked by said total weights in step 1311. Then, in step 1312, the lexical chains with the largest total weights are referenced back to their parent question-answer pairs, and the question-answer pair is outputted for display or processing. The process then ends. In this way, the interactive question returns the FAQ entry or entries that are most relevant to the inquiry.
  • FIG. 13B shows a flow chart illustrating the process for another embodiment of determining semantic similarities for semantic searches shown generally at 1303B. This process is similar to the process outlined in FIG. 13A, but requires less computational resources and may produce minor variation in results. The beginning of this process is intended to be from step 1202, of FIG. 12. However, this process 1203B is not limited to such context, and may be employed in many suitable semantic similarity comparisons. At step 1320 a placeholder T is set to a base value of 1. Then at step 1321 the placeholder T is compared to the number of lexical chains that were generated in step 1109 of FIG. 11 for the question-answer pairs of the FAQ. If T is greater that the total number of lexical chains, then the process proceeds to step 1329 where lexical chains are ranked by total weights. This ensures that all lexical chains generated for the FAQ are compared prior to the ranking of semantic similarities.
  • Otherwise, if, at step 1321, T is less than the total number lexical chains of the FAQ, a second placeholder U is set to 1 at step 1322. Then at step 1323 the placeholder U is compared to the number of elements in lexical chain T. If U is greater than the total elements in the lexical chain T then the weight of the lexical chain T is multiplied by a modifier ZT at step 1327. Modifier ZT provides for corrections to be made for differences between lexical chains, such as for chain length. For example, a highly refined question-answer pair will tend to produce a relatively short lexical chain, since the subject matter explored by the FAQ entry is refined and specific. However, a broad question and answer pair may result in a very long lexical chain. In some circumstances, the interactive question may have more aggregate similarities with the long lexical chain due to its broad nature; however the interactive question is better answered by the narrower, in-depth FAQ entry. In such an instance the modifier ZT may be applied to increase the weight of the shorter lexical chain's semantic similarities.
  • Then a value of 1 is added to T in step 1328. This ensures that all elements within the lexical chain T are compared to the interactive question. The process then returns to step 1321, where the increased placeholder T is compared to the number of lexical chains for the FAQ.
  • Else, if, at step 1323, U is less than the total elements in the lexical chain T, then, at step 1324, a query is made if the element U of lexical chain T is matched by any element in the interactive question. If element U matches an element in the question, then 1 is added to the total weight of the lexical chain T in step 1325. Then, in step 1326, a value of 1 is added to U. This ensures that all elements U within the lexical chain T are compared to the interactive question. The process then returns to step 1323, where the increased placeholder U is compared to the number of elements of the lexical chain T.
  • Else, if at step 1324, element U does not match any element in the question, then a value of 1 is added to U at step 1326. This ensures that all elements U within the lexical chain T are compared to the interactive question. The process then returns to step 1323, where the increased placeholder U is compared to the number of elements of the lexical chain T. In this manner the total weight of the lexical chain T is the sum of the direct matches of its elements and the question.
  • After all lexical chains have been processed for their total weights; the lexical chains may be ranked by said total weights in step 1329. Then, in step 1330, the lexical chains with the largest total weights are referenced back to their parent question-answer pairs, and the question-answer pair is outputted for display or processing. The process then ends. In this way, the interactive question returns the FAQ entry or entries that are most relevant to the inquiry.
  • FIG. 14 shows a flow chart illustrating the process for an embodiment of a general semantic search utilizing the natural language processor system, shown generally at 1400. Process 1400 is a more general semantic searching method then those embodiments disclosed in FIGS. 11 to 13B. For example, process 1400 does not require a FAQ style data set, but rather answer data in the form of a plurality of answer entries. These answer entries may include separate data sources, previous dialogue conversations, news articles, press releases, or virtually any series of informational text.
  • The process begins by parallel process of Step 1401 and 1402. Either step 1401 or 1402 may be performed first, or they may be performed simultaneously. At step 1401 a question language is received. At step 310 the language question is processed through natural language analysis as illustrated in FIG. 4. At step 1402, answer entries are queried. Said answer entries are analyzed through the Natural Language Analyzer 130 for semantics at step 310. At step 1407 the semantics for each answer entry is compared for relatedness to the results the semantics for the question determined at steps 310. At step 1408, the results of the relatedness between the semantics of each answer entry and the question are outputted for further processing. The process then ends.
  • FIG. 15A shows a flow chart illustrating the process for one embodiment of determining semantic similarities for general semantic searches utilizing the Autonomous Avatar Driver System 100 shown generally at 1407A. The beginning of this process is intended to be from step 1406, of FIG. 14. However, this process 1407A is not limited to such context, and may be employed in any suitable semantic similarity comparison. At step 1501 a placeholder Q is set to a base value of 1. Then at step 1502 the placeholder Q is compared to the number of answer entries in the answer data set. If Q is greater that the total number of answer entries then the process proceeds to step 1510 where answer entries are ranked by total weights. This ensures that all answer entries are compared prior to the ranking of semantic similarities.
  • Otherwise, if, at step 1502, Q is less than the total number answer entries, a second placeholder R is set to 1 at step 1503. Then at step 1504 the placeholder R is compared to the number of elements in lexical chain of answer entry Q. If R is greater than the total elements in the lexical chain of answer entry Q then the weight of the lexical chain of answer entry Q is multiplied by a modifier ZQ at step 1508. Modifier ZQ provides for corrections to be made for differences between lexical chains, such as for chain length. For example, a highly refined answer entry will tend to produce a relatively short lexical chain. However, a broad answer entry may result in a very long lexical chain. In some circumstances, the interactive question may have more aggregate similarities with the long lexical chain due to its broad nature; however the interactive question is better answered by the narrower, in-depth answer entry. In such an instance the modifier ZQ may be applied to increase the weight of the shorter lexical chain's semantic similarities.
  • Then a value of 1 is added to Q in step 1509. This ensures that all elements within the lexical chain of answer entry Q are compared to the semantics of the question. The process then returns to step 1502, where the increased placeholder Q is compared to the number of answer entries.
  • Else, if, at step 1504, R is less than the total elements in the lexical chain of answer entry Q, then, at step 1505, a weight is assigned to element R by comparing the semantics of the question to element R. A direct semantic match may generate a higher weight value to element R than a more attenuated semantic relationship. At step 1506 the weight assigned to element R is added to the total weight of answer entry Q. Then, at step 1507, a value of 1 is added to R. The process then returns to step 1504, where the increased placeholder R is compared to the number of elements of the lexical chain of answer entry Q. In this manner the total weight of the lexical chain of answer entry Q is the sum of the weights of its respective elements.
  • After all lexical chains have been processed for their total weights; the answer entries may be ranked by said total weights in step 1510. The process then ends by proceeding to step 1408 in FIG. 14.
  • FIG. 15B shows a flow chart illustrating the process for another embodiment of determining semantic similarities for general semantic searches utilizing the Autonomous Avatar Driver System 100, shown generally at 1407B. This process is similar to the process outlined in FIG. 15A, but requires less computational resources and may produce minor variation in results. The beginning of this process is intended to be from step 1406, of FIG. 14. However, this process 1407B is not limited to such context, and may be employed in any suitable semantic similarity comparison. At step 1521 a placeholder Q is set to a base value of 1. Then at step 1522 the placeholder Q is compared to the number of answer entries. If Q is greater that the total number of answer entries then the process proceeds to step 1530 where answer entries are ranked by total weights. This ensures that all answer entries are compared prior to the ranking of semantic similarities.
  • Otherwise, if, at step 1522, Q is less than the total number answer entries, a second placeholder R is set to 1 at step 1523. Then at step 1524 the placeholder R is compared to the number of elements in lexical chain of answer entry Q. If R is greater than the total elements in the lexical chain of answer entry Q then the weight of the lexical chain of answer entry Q is multiplied by a modifier ZQ at step 1528. Modifier ZQ provides for corrections to be made for differences between lexical chains, such as for chain length. For example, a highly refined answer entry will tend to produce a relatively short lexical chain. However, a broad answer entry may result in a very long lexical chain. In some circumstances, the question may have more aggregate similarities with the long lexical chain due to its broad nature; however the interactive question is better answered by the narrower, in-depth answer entry. In such an instance the modifier ZQ may be applied to increase the weight of the shorter lexical chain's semantic similarities.
  • Then a value of 1 is added to Q in step 1529. This ensures that all elements within the lexical chain of answer entry Q are compared to the question. The process then returns to step 1522, where the increased placeholder Q is compared to the number of answer entries.
  • Else, if, at step 1524, R is less than the total elements in the lexical chain of answer entry Q, then, at step 1525, a query is made if the element R of lexical chain of answer entry Q is matched by any element in the question. If element R matches an element in the question, then 1 is added to the total weight of answer entry Q in step 1526. Then, in step 1527, a value of 1 is added to R. This ensures that all elements R within the lexical chain of answer entry Q are compared to the question. The process then returns to step 1524, where the increased placeholder R is compared to the number of elements of the lexical chain of answer entry Q.
  • Else, if at step 1525, element R does not match any element in the question, then a value of 1 is added to R at step 1527. This ensures that all elements R within the lexical chain of answer entry Q are compared to the question. The process then returns to step 1524, where the increased placeholder R is compared to the number of elements of the lexical chain of answer entry Q. In this manner the total weight of answer entry Q is the sum of the direct matches of its lexical chain's elements and the question.
  • After all answer entries have been processed for their total weights; the answer entries may be ranked by said total weights in step 1530. The process then ends by proceeding to step 1408 in FIG. 14.
  • FIG. 16 shows a flow chart illustrating the process for generating tag clouds shown generally at 1600. Tag clouds are a collection of keywords that may be used in order to enable the user to interact more efficiently with a user interface. Tag clouds are well known by those skilled in the art; however, the usage of tag clouds that are responsive to natural language processing is unique to the present invention. This embodiment of the present invention deviates from the semantic searching embodiments; however, this embodiment is intended to illustrate another useful application of the Autonomous Avatar Driver System 100.
  • The process begins at step 1601, where conversation text data is queried for downstream semantic processing. Beyond conversation data, additional data may be available for query, such as conversation histories, databases and external data sources. In this way the avatar may be able to generate tag clouds that provide somewhat impulsiveness conversation developments. Such impulsiveness allows more avenues of conversation to develop, thereby enhancing the users' experience. Additionally, impulsiveness is a decidedly human trait, thus adding to the believability of the avatar. For example, in a gaming context if the avatar is assigned a role as a farmer, and the user engages in language regarding the weather, the tag cloud may provide keywords about weather, weather predictions and weather's effects upon crops. This is made possible by supplying the avatar with knowledge templates regarding farming practices, as is discussed previously.
  • At step 523 sentence detection is performed on the queried text. Then, in step 524, the data is tokenized, thereby cleaning the text by removing special characters. Then, in step 430, the cleansed text is tagged as is discusses above at Table 1. The tagged text is then parsed, or chunked, in step 440, as was seen at Table 2. Then at step 1606, the nouns and verbs are extracted from the text in order to generate lexical chains at step 1607. The lexical chains are then filtered at step 1608 to generate tag clouds. At 1609 the tag clouds are generated.
  • The filtering, at step 1608, may filter individual chains, or filter multiple chains. For instance, the available lexical chains may be compared for similar content. Lexical chains that are too similar may be synthesized, or one of them may be eliminated. This reduces the redundancy of lexical chains used in the construction of the tag clouds. For individual lexical chains, it may be advantageous to reduce each chain to a predetermined number of elements. Redundant elements may be synthesized or eliminated first. Among strings of hyponyms and hypernyms an element with a medium level of detail may be retained and the remaining hyponyms and hypemyms may be discarded. Again, the purpose of such filtering is the reduction of the redundant elements used in the construction of the tag clouds.
  • The generation of the tag clouds may be as simple as the displaying of the remaining elements of the lexical chains. Alternatively, intact chains may be displayed. Moreover, in some embodiments, the tag clouds could even include dialog from the Natural Language Generator 140, which may utilize the lexical chains generated in step 1607 to generate natural language “thoughts” or conversation suggestions in natural language.
  • IV. Autonomous Avatar Driver Systems and Methods
  • FIG. 17 shows a schematic block diagram illustrating the Autonomous Avatar Driver 150. The Autonomous Avatar Driver 150 includes a Sourcer 1711, a Segmenter 1712, a Parser 1713, a Semantic Engine 1714, an Analytics Engine 1715, a Pragmatics Analyzer 1716, a Recommender 1717 and a Response Generator 1718. Each component of the Autonomous Avatar Driver 150 may be coupled with one another, thereby enabling communication between the respective components of the Autonomous Avatar Driver 150. It should be noted that additional, or fewer, components may be included in the Autonomous Avatar Driver 150 as is desired. The illustrated Autonomous Avatar Driver 150 is intended to be purely exemplary in nature. In some embodiments, the Autonomous Avatar Driver 150 may utilize the Natural Language Processor 190 for many of the analytic and semantic analysis of the avatar dialog.
  • The exemplary Data Source 102 a and Language Source 103 a may provide input to the Autonomous Avatar Driver 150 via the Sourcer 1711. The Segmenter 1712 may segment the language, and the Parser 1713 may parse the language. The Semantic Engine 1714 may receive the segmented and parsed language and perform a semantic analysis. Details of these processes, as performed for these particular embodiments of the Autonomous Avatar Driver 150 are further explained below. Of course alternate methods of segmenting, parsing and semantic analysis of the received language may be utilized. For example the generic Natural Language Processor 190 may segment, parse and analyze for the received language semantics as previously disclosed.
  • The Analytics Engine 1715 performs analytic analysis to the semantically analyzed language. Analytics may include filtering for common words, token frame statistics and N-gram analysis. The Pragmatics Analyzer 1716 may take the language and resolve pragmatic issues, such as slang usage, knowledge template application, grouping of proper nouns, and estimation of affect.
  • The Recommender 1717 provides recommendations from the analyzed language to the Data Consumer 125. These recommendations may include highly relevant contextual advertising, statistical analysis, and improvement of the user's experience through user recommendations. For example, if the dialog is regarding music, the analyzed language could identify the user's preference of bands. These preferences may then be compared to statistical databases of similar user's preferences to provide music that is likely to be enjoyed by the user. Such “product recommendation” is not new by itself, however by intelligently and contextually analyzing a dialog the Autonomous Avatar Driver 150 may provide highly relevant, accurate and useful recommendations.
  • Likewise the Recommender 1717 may provide recommendations within a social network environment of other individuals that have similar interests, and/or compatible personalities. This may be of particular use in casual social networks such as Facebook®, as well dating style social networks such as match.com®.
  • The Response Generator 1718 may utilize the analyzed language to generate relevant responses that may in turn drive the Autonomous Avatar 1720. Such response information may include graphical information, such as avatar expression, as well as relevant dialog. In some embodiments, the Response Generator 1718 may, additionally, receive recommendation information from the Recommender 1717 to incorporate the generated recommendation within the response. Therefore, new and relevant topics, tailored to the conversation, may be included in the response. Additionally, Autonomous Avatar 1720 may provide relevant and contextual advertising when appropriate.
  • FIG. 18 shows a schematic block diagram illustrating the Sourcer 1711 of the Autonomous Avatar Driver 150. The Sourcer 1711 may include a Dialog Manager 1802 and a Source Crawler 1804. The Dialog Manager 1802 may receive dialog information from the Language Source 103 p. Of course the Sourcer 1711 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150. The Source Crawler 1804 may, in some embodiments, provide a more active role of crawling the Data Source 102 a for information as pertains to the dialog received by the Dialog Manager 1802. The Dialog Manager 1802 may be coupled to the Source Crawler 1804 to facilitate the searching for relevant information by the Source Crawler 1804. The Source Crawler 1804 may utilize any crawling technology, as is well known by those skilled in the art.
  • The Source Crawler 1804 may search any appropriate Data Source 102 a; however, social media, and structured data sources, such as Wikipedia, may be routinely sourced due to the ease of searching, as well as the utility of the data gathered for most interactions. For example, the format of typical web-logs, or blogs, is very similar to a screenplay. Screenplays and blogs typically have topical formats, which are highly useful for gathering additional contextual information.
  • FIG. 19 shows a schematic block diagram illustrating the Segmenter 1712 of the Autonomous Avatar Driver 150. The Segmenter 1712 includes a Particle Representor 1902, a Character Converter 1904 and a Splitter 1906, each couple to one another. Of course the Segmenter 1712 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150. The Particle Representor 1902 may separate the dialog into “particles”, or discrete logical subsections of the dialog. For example, “particles” of a screenplay would include the individual acts and scenes. This particle representation may then represent the language in a data structure format. Particle representation may additionally include speaker detection.
  • The Character Converter 1904 may convert the dialog characters. In some embodiments, the Character Converter 1904 will convert the dialog to UTF-8 encoding. However, additional conversion formats may be utilized by the Character Converter 1904 as is known by those skilled in the art.
  • The Splitter 1906 may split the sentences of the dialog. In some embodiments, a max entropy splitter may be utilized for the Splitter 1906.
  • FIG. 20 shows a schematic block diagram illustrating the Parser 1713 of the Autonomous Avatar Driver 150. The Parser 1713 may include a Grammar Linker 2002, a Grammatical Mood Analyzer 2004, a Tagger 2006 and a Variant Pruner 2008, each coupled to one another. Of course the Parser 1713 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150. The Grammar Linker 2002 may link grammatically useful word combinations. These combinations provide precision of the language when transformed into computer language. An example of words that may be linked includes “have” and “being”. These word combinations denote specific grammatical stances when paired and conjugated.
  • The Grammatical Mood Analyzer 2004 may determine the grammatical mood of the language. The grammatical moods include interrogative, declarative and imperative. See table 3 above. Of course additional moods may be determined, such as energetic and conditional, as is desired for functionality.
  • The Tagger 2006 may tag the words of the dialog, as is discussed above in the natural language processor section. Tagging the language is fundamental to the generation of lexical chains, as well as the ability for the Autonomous Avatar Driver 150 to derive accurate semantic information from the language. In some embodiments, a max entropy tagger may be utilized for the Tagger 2006.
  • The Variant Pruner 2008 may prune variants within the dialog by utilizing Bayesian statistics. Results from the Variant Pruner 2008 may be output to the NLP Servicer 2020. The NLP Servicer 2020 may include any entity that requires the parsed language. Examples of the NLP Servicer 2020 include chat rooms or other data compilers.
  • FIG. 21 shows a schematic block diagram illustrating the Semantic Engine 1714 of the Autonomous Avatar Driver 150. The Semantic Engine 1714 may include a Token Framer 2102, a Lemmatizer 2104, a Semantic Network Generator 2106 and a Co-reference Resolver 2108, each coupled to one another. Of course the Semantic Engine 1714 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150. The Token Framer 2102 may tokenize the dialog as discussed previously. The Lemmatizer 2104 may lemmatize the dialog, wherein the verbs and nouns are reduced back into basic forms. The Semantic Network Generator 2106 may generate meaningful graphs of the concepts, wherein relationships of the concepts are defined. Statistical profiling may be utilized in the generation of the semantic networks.
  • Such semantic networks may require significant processing power and time to complete. Thus, in some embodiments, corpuses may be utilized to pre-compute semantic networks, which may be updated as required. These pre-computed semantic networks may then be referenced when the analyzed language contains nodes of commonality. Thus processing time may be reduced for accurate semantic network generation.
  • Lastly, the Co-reference Resolver 2108 may resolve co-referencing ambiguities within the dialog. The Co-reference Resolver 2108 may resolve the co-references via pattern matching, context appropriateness and by utilizing verb reference to a subject.
  • FIG. 22 shows a schematic block diagram illustrating the Analytics Engine 1715 of the Autonomous Avatar Driver 150. The Analytics Engine 1715 may include a Common Word Filter 2202, a Token Frame Statistics Analyzer 2204, and an N-gram Analyzer 2206, each coupled to one another. Of course the Analytics Engine 1715 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150. The Common Word Filter 2202 may filter the dialog for common words, thereby reducing redundancy.
  • The Token Frame Statistics Analyzer 2204 may sum lemmatized dialog. Such a count is akin to “keyword hits” utilized by standard searching systems; however by counting the lemmatized dialog, more contextually accurate searches are possible. Additionally, the Token Frame Statistics Analyzer 2204 may also perform additional statistical analysis on the dialog as is desired for system functionality. An example would include nodal analysis of the semantic network.
  • The N-gram Analyzer 2206 may perform an N-gram analysis on the dialog. In the English language, typically 3 or 4 words may be utilized for the N-gram analysis. Of course different word numbers may be utilized as is desired. Additionally, the number of words utilized for the N-gram analysis is highly language dependent. In an N-gram analysis, a sliding window along N words may be utilized to generate popularity statistics.
  • FIG. 23 shows a schematic block diagram illustrating the Pragmatics Analyzer 1716 of the Autonomous Avatar Driver 150. The Pragmatics Analyzer 1716 may include a Slang Resolver 2302, a Knowledge Template Resolver 2304, a Proper Noun Grouper 2306 and an Affect Estimator 2308, each coupled to one another. Of course the Pragmatics Analyzer 1716 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150. The Pragmatics Analyzer 1716 enables the Autonomous Avatar Driver 150 to filter certain behaviors of human users that make contextual analysis of the dialog problematic.
  • The Slang Resolver 2302 may resolve slang language utilizing statistical collection of word usage as gained from ongoing corpuses, such as blogs. These slang terms may then populate a slang database for rapid retrieval when resolving slang in the dialog.
  • The Knowledge Template Resolver 2304 may provide knowledge representation via knowledge templates as disclosed above in the discussion of the natural language processing. The Proper Noun Grouper 2306 may group together proper nouns as identified by grammatical form. The Affect Estimator 2308 may estimate dialog affect, or emotion, by analyzing the lemmatized dialog in conjunction with the grammatical mood of the dialog.
  • FIG. 24 shows a schematic block diagram illustrating the Recommender 1717 of the Autonomous Avatar Driver 150. The Recommender 1717 may include a Cloud Aggregator 2402, a Hood Clusteror 2404, a Social Network Recommender 2406 and a Contextual Advertising Generator 2408, each coupled to one another. Of course the Recommender 1717 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150. Each component of the Recommender 1717 may output their recommendations to the Data Consumer 125.
  • The Cloud Aggregator 2402 may aggregate the tag clouds of related dialog. The Hood Clusteror 2404 may cluster people by their closest distance by interest or compatibility. This may have particular utility for “people fit” services on social networks and dating services. The Social Network Recommender 2406 may recommend contacts on social networks. This may be of particular use for business to business contact generation. Lastly, the Contextual Advertising Generator 2408 may generate highly contextual and highly targeted advertising. Such advertising recommendations are only possible by analyzing not only content, but also meaning of the dialog.
  • Additionally, it should be noted that the Recommender 1717 may assist the Response Generator 1718 to provide dialog for the Autonomous Avatar 1720 that is not only responsive, but also that may include recommendations that are relevant to the user.
  • FIG. 25 shows a schematic block diagram illustrating the Response Generator 1718 of the Autonomous Avatar Driver 150. The Response Generator 1718 may include a Summarizer 2502 and a Dialog Manager 2504 coupled to one another. Of course the Response Generator 1718 may include more, or fewer, components as is required for the desired functionality of the Autonomous Avatar Driver 150. The Summarizer 2502 may summarize the response that is generated. The Dialog Manager 2504 may then manage the dialog for the Autonomous Avatar 1720. It should be noted that while the Response Generator 1718 is discussed as generating dialog, additional responses may be generated by the Response Generator 1718, including but not limited to graphical emotional expressions, movements and non-word exclamations.
  • FIG. 26A shows a schematic block diagram of the Virtual Universe 2600 for the autonomous avatar driver system of FIG. 17. The Virtual Universe 2600 may be broken down into five subcategories: Virtual Overlays of Real World Data 2601, WEB Communities 2602, Massively Multiplayer Online Games (MMOGs) 2603, Social Worlds 2604, and Telecom 2605. Examples of Virtual Overlays of Real World Data 2601 include, but are not limited to, Google Earth and Microsoft Flight Simulator X. Examples of WEB Communities 2602 include, but are not limited to, YouTube and MySpace. Examples of MMOGs 2603 include, but are not limited to, World of Warcraft, Guild Wars and Hive. Examples of Social Worlds 2604 include, but are not limited to, Second Life and Neopets. Examples of Telecom 2605 include, but are not limited to, cell phones, BlackBerry Devices and Personal Digital Assistants (PDA). Additional subcategories may exist, or may emerge with new technology. It is intended that these additional subcategories be incorporated into the Virtual Universe 2600. The Autonomous Avatar Driver 150 is coupled to the subcategories of the Virtual Universe 2600 through the WAN 101.
  • A logical block diagram of the Virtual Universe 2600 is shown in FIG. 26B. Each Virtual Environments 2611 a, 2611 b to 2611 x, 2612 a, 2612 b to 2612 y, 2613 a, 2613 b to 2613 z, 2614 a, 2614 b to 2614 m, 2615 a, 2615 b to 2615 n is coupled to the WAN 101. Each subcategory, Virtual Overlays of Real World Data 2601, WEB Communities 2602, MMOGs 2603, and Social Worlds 2604, and Telecom 2605, may include multiple Virtual Environments 2611 a to 2615 n. Moreover, some Virtual Environments 2611 a to 2615 n may be hybrids of these subcategories. Thus, while the line between specific subcategories may become increasingly indistinct, the boundaries between individual Virtual Environments 2611 a to 2615 n are distinct and nearly impassable. Occasionally, the Virtual Overlays of Real World Data 2601 have provided some connectivity between Virtual Environments 2611 a to 2615 n as shown in FIG. 25; however this connectivity is limited in scope. The Autonomous Avatar Driver 150, on the other hand, is able to access all the Virtual Environments 2611 a to 2615 n thereby enabling the driving of the Autonomous Avatar 1720 in any Virtual Environment 2611 b.
  • A logical block diagram of an exemplary Virtual Environment 2611 b is shown in FIG. 26C. Within each Virtual Environment 2611 b exists an Enabler 2631. The Enabler 2631 allows for Autonomous Avatar 1720 a, 1720 b to 1720 t to access the WAN 101, and eventually the Autonomous Avatar Driver 150. In some embodiments, each Virtual Environment 2611 b has a corresponding Enabler 2631. However, any number of Autonomous Avatars 1720 a to 1720 t may exist within a Virtual Environment 2611 b at any given time.
  • A logical block diagram of an exemplary Autonomous Avatar 1720 a is shown in FIG. 26D. In some embodiments, Autonomous Avatar 1720 a may include Physical Attributes 2641, Intellectual Attributes 2642 and Emotional Attributes 2643. Each attribute may be coupled to an Avatar Engine 2644 which may coordinate and manage each attribute. Physical Attributes 2641 may include Autonomous Avatar's 1720 a physical statistics, such as strength, and appearance data. Intellectual Attributes 2642 may include the Autonomous Avatar's 1720 a backstory, history and memory. Emotional Attributes 2643 may include the Autonomous Avatar's 1720 a emotional disposition, and reaction and response algorithms.
  • FIG. 27 shows a flow chart illustrating the process for driving the Autonomous Avatar 1720, shown generally at 2700. The process begins from step 2702 where the dialog is received by Dialog Manager 1802. Typically a user, driving a personal avatar may interact with the Autonomous Avatar 1720, thereby generating the dialog. However, it is intended that any Language Source 103 p may provide the dialog.
  • Then, at step 2604 the data source is crawled by the Source Crawler 1804. The process then proceeds to step 2706 where the received dialog is segmented by utilizing the Segmenter 1712. Then, at step 2608 the dialog is parsed by utilizing the Parser 1713. The process then proceeds to step 2710 where semantic analysis is performed by utilizing the Semantic Engine 1714. At 2712 analytic analysis is performed by utilizing the Analytics Engine 1715. Pragmatic analysis is then performed by utilizing the Pragmatics Analyzer 1716, at step 2714. The process then proceeds to step 2716 where recommendations are generated by utilizing the Recommender 1717. Lastly, at step 2718, dialog is generated for the Autonomous Avatar 1720 by utilizing the Response Generator 1718. The process then ends.
  • FIG. 28 shows a flow chart illustrating the process for dialog segmentation, shown generally at 2706. The process begins from step 2704 of FIG. 27. The process then proceeds to step 2802 where particle representation is performed. Particle representation includes organizing the dialog into a data structure by utilizing the Particle Representor 1902.
  • The process then proceeds to step 2804 where character conversion is performed. Character conversion may utilize the Character Converter 1904. The process then proceeds to step 2806 where the dialog is split into individual sentences. The Splitter 1906 may be utilized for dialog splitting. The process then concludes by proceeding to step 2708 of FIG. 27.
  • FIG. 29 shows a flow chart illustrating the process for dialog parsing, shown generally at 2708. The process begins from step 2706 of FIG. 27. The process then proceeds to step 2902 where grammar is linked. As previously discussed, grammar linking of particular word groups, such as “have” and “being” may be particularly useful for providing precision of the dialog when converted to computer language. Grammatical linking may utilize the Grammar Linker 2002.
  • Then at step 2904 grammatical mood is determined by utilizing the Grammatical Mood Analyzer 2004. The process then proceeds to step 2906 where tagging of the dialog is performed. As previously mentioned, tagging of the dialog provides an abundance of information as to the sentence structure, and ultimately to the meaning of the dialog. Tagging may be performed by the Tagger 2006. Lastly, at step 2908 variants are pruned by utilizing the Variant Pruner 2008. Variant pruning may utilize Bayesian statistics or other statistical methodologies. The process then concludes by proceeding to step 2710 of FIG. 27.
  • FIG. 30 shows a flow chart illustrating the process for semantic analysis, shown generally at 2710. The process begins from step 2708 of FIG. 27. The process then proceeds to step 3002 where the dialog is tokenized by utilizing a token frame dictionary lookup. The tokenization may utilize the Token Framer 2102.
  • The process then proceeds to step 3004 where semantic lexicons are generated by utilizing the Lemmatizer 2104. A detailed discussion of the generation of semantic lexicons is provided above in the discussion of natural language processing.
  • Then, at step 3006 semantic networking may be performed, by utilizing the Semantic Network Generator 2106. As previously noted, prototypical semantic networks may be pre-computed for access later by commonalities to the dialog, thereby reducing the processing demands on the Autonomous Avatar Driver 150.
  • The process then proceeds to step 3008 where co-references are resolved by utilizing the Co-reference Resolver 2108. Co-references may be resolved in three ways: pattern matching, context appropriateness, and verb reference to a subject. The process then concludes by proceeding to step 2712 of FIG. 27.
  • FIG. 31 shows a flow chart illustrating the process for analytic analysis, shown generally at 2712. The process begins from step 2710 of FIG. 27. The process then proceeds to step 3102 where common words are filtered for by utilizing the Common Word Filter 2202. Common words may include such regularly used words as “the” and “a”. The process then proceeds to step 3104 where token frame statistics are performed by utilizing the Token Frame Statistics Analyzer 2204. Token frame statistics may include counting lemmatized words and nodal analysis. However, additional statistical inquiries may be preformed as is desired for Autonomous Avatar Driver 150 functionality.
  • The process then proceeds to step 3106 where the N-gram analysis is performed by utilizing the N-gram Analyzer 2206. As previously noted, the N-gram analysis includes sliding a window of N words along the dialog to generate statistics on word popularity. The number of N words may vary dependent upon language of the dialog. The process then concludes by proceeding to step 2714 of FIG. 27.
  • FIG. 32 shows a flow chart illustrating the process for pragmatic analysis, shown generally at 2714. The process begins from step 2712 of FIG. 27. The process then proceeds to step 3202 where slang is resolved by utilizing the Slang Resolver 2302. Then, at step 3204 knowledge may be represented using knowledge templates. The Knowledge Template Resolver 2304 may be utilized to generate the knowledge templates. The process then proceeds to step 3206 where proper nouns may be grouped by utilizing the Proper Noun Grouper 2306. Lastly, at step 3208 affect may be estimated using a combination of grammatical mood analysis and contextual referencing. Affect estimation may utilize the Affect Estimator 2308. The process then concludes by proceeding to step 2716 of FIG. 27.
  • FIG. 33 shows a flow chart illustrating the process for recommendation generation, shown generally at 2716. The process begins from step 2714 of FIG. 27. The process then proceeds to step 3302 where tag clouds are aggregated by utilizing the Cloud Aggregator 2402. Then, at step 3304, individuals may be clustered into “neighborhoods” by related interest or compatibility. Additional clustering criteria may also be utilized, such as abilities, political ideologies or diversity. Clustering may utilize the Hood Clusteror 2404.
  • The process then proceeds to step 3306 where individuals may be recommended to one another across a social network by utilizing the Social Network Recommender 2406. Lastly, at step 3308, contextual advertising may be recommended by utilizing the Contextual Advertising Generator 2408. The process then concludes by proceeding to step 2718 of FIG. 27.
  • FIG. 34 shows a flow chart illustrating the process for avatar dialog generation, shown generally at 2718. The process begins from step 2716 of FIG. 27. The process then proceeds to step 3402 where dialog is summarized by utilizing the Summarizer 2502. This dialog may then be managed at step 3404 by utilizing the Dialog Manager 2504. The process then ends.
  • In sum, the present invention provides an Autonomous Avatar Driver System 100 that provides for advanced dialog management that is highly contextual, and which provides highly relevant recommendation based upon conversation dialog. Such autonomous avatar drivers are useful for driving believable avatars in conjunction with massively multiplayer Online Games, virtual social worlds, online web communities corporate web sites, health care and educational programs and websites. Autonomous avatar driver systems may be entirely software, entirely hardware, or a combination of software and hardware. The advantages of such an efficient system include enhancing the believability of avatars, having knowledge representation from vast data sources, improved searching which utilizes semantics, improved tag clouds, improved corporate relations, advanced recommendation abilities and positive repercussions for the healthcare and education industries.
  • While this invention has been described in terms of several preferred embodiments, there are alterations, modifications, permutations, and substitute equivalents, which fall within the scope of this invention. Although sub-section titles have been provided to aid in the description of the invention, these titles are merely illustrative and are not intended to limit the scope of the present invention.
  • It should also be noted that there are many alternative ways of implementing the methods and apparatuses of the present invention. It is therefore intended that the following appended claims be interpreted as including all such alterations, modifications, permutations, and substitute equivalents as fall within the true spirit and scope of the present invention.

Claims (20)

1. A computer implemented method for driving an autonomous avatar, useful in association with at least on language source, the method comprising:
receiving dialog from the language source;
segmenting the dialog;
parsing the segmented dialog;
analyzing the parsed dialog for semantics;
analyzing the semantically analyzed dialog for analytics;
analyzing the analytically analyzed dialog for pragmatics;
generating recommendations from the pragmatically analyzed dialog; and
generating responses for the autonomous avatar using the pragmatically analyzed dialog.
2. The computer implemented method, as recited in claim 1, further comprising receiving external data from at least one data source.
3. The computer implemented method, as recited in claim 2, wherein the segmenting the dialog includes at least one of performing particle representation, converting characters and splitting dialog.
4. The computer implemented method, as recited in claim 2, wherein the parsing the segmented dialog includes at least one of linking grammar, generating grammatical mood, tagging the dialog and pruning variants.
5. The computer implemented method, as recited in claim 2, wherein the analyzing the parsed dialog for semantics includes at least one of lookup of token frame, generation of semantic lexicons, generation of semantic networks and resolving co-reference ambiguity.
6. The computer implemented method, as recited in claim 2, wherein the analyzing the semantically analyzed dialog for analytics includes at least one of filtering for common words, generating token frame statistics and analyzing N-grams for word popularity.
7. The computer implemented method, as recited in claim 6, wherein the generating token frame statistics includes at least one of counting lemmatized words, and nodal analysis.
8. The computer implemented method, as recited in claim 2, wherein the analyzing the analytically analyzed dialog for pragmatics includes at least one of resolving slang, generating knowledge templates, grouping proper nouns and estimating affect.
9. The computer implemented method, as recited in claim 2, wherein the generating recommendations from the pragmatically analyzed dialog includes at least one of generating aggregate tag clouds, clustering the language sources into neighborhoods, recommending social networking, and generating contextual advertising.
10. The computer implemented method, as recited in claim 2, wherein the generating responses for the autonomous avatar also includes incorporating the generated recommendations.
11. An automated avatar driver system useful in association with an avatar and at least one language source, the automated avatar driver system comprising:
a sourcer configured to receive dialog from the language source;
a segmentor configured to segment the dialog;
a parser configured to parse the segmented dialog;
a semantic engine configured to analyze the parsed dialog for semantics;
an analytics engine configured to analyze the semantically analyzed dialog for analytics;
a pragmatics analyzer configured to analyze the analytically analyzed dialog for pragmatics;
a recommender configured to generate recommendations from the pragmatically analyzed dialog; and
a response generator configured to generate responses for the autonomous avatar using the pragmatically analyzed dialog.
12. The automated avatar driver system of claim 11, wherein the sourcer is further configured to receive external data from at least one data source.
13. The automated avatar driver system of claim 12, wherein the segmentor includes at least one of a particle representor configured to perform particle representation, a character converter configured to convert characters and a splitter configured to split dialog.
14. The automated avatar driver system of claim 12, wherein the parser includes at least one of a linker configured to link grammar, a grammatical mood analyzer configured to generate grammatical mood, a tagger configured to tag the dialog and a variant pruner configured to prune variants of the dialog.
15. The automated avatar driver system of claim 12, wherein the semantics engine includes at least one of a token framer configured to lookup token frames, a lemmatizer configured to generate semantic lexicons, a semantic network generator configured to generate semantic networks and a co-reference resolver configured to resolve co-reference ambiguity.
16. The automated avatar driver system of claim 12, wherein the analytics engine includes at least one of a common word filter configured to filter dialog for common words, a token frame statistics engine, and an N-gram analyzer configured to analyze N-grams for word popularity.
17. The automated avatar driver system of claim 16, wherein the token frame statistics engine configured to count lemmatized words, and perform nodal analysis.
18. The automated avatar driver system of claim 12, wherein the pragmatics analyzer includes at least one of a slang resolver configured to resolve slang, a knowledge template resolver configured to generate knowledge templates, a proper noun grouper configured to group proper nouns and an affect estimator configured to estimate affect of the dialog.
19. The automated avatar driver system of claim 12, wherein the recommender includes at least one of a cloud aggregator configured to generate aggregate tag clouds, a hood cluster or configured to cluster the language sources into neighborhoods, a social network recommender configured to recommend social networking, and a contextual advertising generator configured to generate contextual advertising.
20. The automated avatar driver system of claim 12, wherein the response generator also includes incorporating the generated recommendations from the recommender.
US11/960,507 2007-03-06 2007-12-19 Systems and methods for an autonomous avatar driver Abandoned US20080221892A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US11/960,507 US20080221892A1 (en) 2007-03-06 2007-12-19 Systems and methods for an autonomous avatar driver
US14/536,626 US10133733B2 (en) 2007-03-06 2014-11-09 Systems and methods for an autonomous avatar driver

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US68281307A 2007-03-06 2007-03-06
US11/960,507 US20080221892A1 (en) 2007-03-06 2007-12-19 Systems and methods for an autonomous avatar driver

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US68281307A Continuation-In-Part 2007-03-06 2007-03-06

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/536,626 Continuation US10133733B2 (en) 2007-03-06 2014-11-09 Systems and methods for an autonomous avatar driver

Publications (1)

Publication Number Publication Date
US20080221892A1 true US20080221892A1 (en) 2008-09-11

Family

ID=39742541

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/960,507 Abandoned US20080221892A1 (en) 2007-03-06 2007-12-19 Systems and methods for an autonomous avatar driver
US14/536,626 Expired - Fee Related US10133733B2 (en) 2007-03-06 2014-11-09 Systems and methods for an autonomous avatar driver

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/536,626 Expired - Fee Related US10133733B2 (en) 2007-03-06 2014-11-09 Systems and methods for an autonomous avatar driver

Country Status (1)

Country Link
US (2) US20080221892A1 (en)

Cited By (80)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050086046A1 (en) * 1999-11-12 2005-04-21 Bennett Ian M. System & method for natural language processing of sentence based queries
US20090030775A1 (en) * 2007-07-26 2009-01-29 Braintexter, Inc. System to generate and set up an advertising campaign based on the insertion of advertising messages within an exchange of messages, and method to operate said system
US20090037822A1 (en) * 2007-07-31 2009-02-05 Qurio Holdings, Inc. Context-aware shared content representations
US20090058862A1 (en) * 2007-08-27 2009-03-05 Finn Peter G Automatic avatar transformation for a virtual universe
US20090186635A1 (en) * 2008-01-22 2009-07-23 Braintexter, Inc. Systems and methods of contextual advertising
US20090210803A1 (en) * 2008-02-15 2009-08-20 International Business Machines Corporation Automatically modifying communications in a virtual universe
US20090210213A1 (en) * 2008-02-15 2009-08-20 International Business Machines Corporation Selecting a language encoding of a static communication in a virtual universe
US20090216625A1 (en) * 2008-02-27 2009-08-27 Adam Jeffrey Erlebacher Systems and Methods for Automated Identification and Evaluation of Brand Integration Opportunities in Scripted Entertainment
US20090216691A1 (en) * 2008-02-27 2009-08-27 Inteliwise Sp Z.O.O. Systems and Methods for Generating and Implementing an Interactive Man-Machine Web Interface Based on Natural Language Processing and Avatar Virtual Agent Based Character
US20100058183A1 (en) * 2008-09-02 2010-03-04 International Business Machines Corporation Method, system, and program product for allocating virtual universe customer service
US20110077934A1 (en) * 2009-09-30 2011-03-31 International Business Machines Corporation Language Translation in an Environment Associated with a Virtual Application
EP2317730A1 (en) * 2009-10-29 2011-05-04 Siemens Enterprise Communications GmbH & Co. KG Method and system to automatically change or update the configuration or setting of a communication system
US20110137844A1 (en) * 2009-12-07 2011-06-09 Alphaport, Inc. Avatar-Based Technical Networking System
US20110213642A1 (en) * 2008-05-21 2011-09-01 The Delfin Project, Inc. Management system for a conversational system
US20110258544A1 (en) * 2010-04-16 2011-10-20 Avaya Inc. System and method for suggesting automated assistants based on a similarity vector in a graphical user interface for managing communication sessions
US20120129612A1 (en) * 2010-11-19 2012-05-24 KABUSHIKI KAISHA SQUARE ENIX (also trading as "SQUARE ENIX CO., LTD.") Game apparatus, game program and information recording medium
US8261307B1 (en) 2007-10-25 2012-09-04 Qurio Holdings, Inc. Wireless multimedia content brokerage service for real time selective content provisioning
US20130218678A1 (en) * 2010-10-19 2013-08-22 Citizennet Inc Systems and methods for selecting and generating targeting information for specific advertisements based upon affinity information obtained from an online social network
US8620850B2 (en) 2010-09-07 2013-12-31 Blackberry Limited Dynamically manipulating an emoticon or avatar
US8628392B1 (en) * 2009-05-22 2014-01-14 Xiaohui Kong Game of actual planning, task/time management, and information sharing
US20140039877A1 (en) * 2012-08-02 2014-02-06 American Express Travel Related Services Company, Inc. Systems and Methods for Semantic Information Retrieval
US20140114998A1 (en) * 2010-11-29 2014-04-24 Viralheat, Inc. Determining demographics based on user interaction
US8738739B2 (en) 2008-05-21 2014-05-27 The Delfin Project, Inc. Automatic message selection with a chatbot
US8769037B2 (en) 2010-11-30 2014-07-01 International Business Machines Corporation Managing tag clouds
US20140278605A1 (en) * 2013-03-15 2014-09-18 Ncr Corporation System and method of completing an activity via an agent
US8868590B1 (en) 2011-11-17 2014-10-21 Sri International Method and system utilizing a personalized user model to develop a search request
US20150006159A1 (en) * 2011-07-28 2015-01-01 Microsoft Corporation Linguistic error detection
US20150004591A1 (en) * 2013-06-27 2015-01-01 DoSomething.Org Device, system, method, and computer-readable medium for providing an educational, text-based interactive game
US20150051910A1 (en) * 2013-08-19 2015-02-19 Nuance Communications, Inc. Unsupervised Clustering of Dialogs Extracted from Released Application Logs
US20150066835A1 (en) * 2013-09-03 2015-03-05 Roger Midmore Computational Tools For Genomic Sequencing and Macromolecular Analysis
US20150066730A1 (en) * 2013-09-03 2015-03-05 Roger Midmore Methods and Systems of Four-Valued Monte Carlo Simulation for Financial Modeling
US20150066476A1 (en) * 2013-09-03 2015-03-05 Roger Midmore Methods and Systems of Four Valued Analogical Transformation Operators Used in Natural Language Processing and Other Applications
WO2015039214A1 (en) * 2013-09-18 2015-03-26 Moldoveanu Mihnea Calin System and method for providing a communication platform with guided dialogical functions
US20150100381A1 (en) * 2013-10-03 2015-04-09 Douglas Petrie Method and System for Increasing the Percentage of Customers Realized from Smart Phone Advertising
US9063927B2 (en) 2011-04-06 2015-06-23 Citizennet Inc. Short message age classification
US20150193415A1 (en) * 2014-01-08 2015-07-09 International Business Machines Corporation Presenting tags of a tag cloud in a more understandable and visually appealing manner
US20150248397A1 (en) * 2014-02-28 2015-09-03 Educational Testing Service Computer-Implemented Systems and Methods for Measuring Discourse Coherence
US9135245B1 (en) * 2012-01-09 2015-09-15 Cox Communications, Inc. Filtering content based on acquiring data associated with language identification
US20150262577A1 (en) * 2013-08-29 2015-09-17 Panasonic Intellectual Property Corporation Of America Speech recognition method and speech recognition apparatus
US9195640B1 (en) 2009-01-12 2015-11-24 Sri International Method and system for finding content having a desired similarity
US20150340026A1 (en) * 2014-05-22 2015-11-26 Palo Alto Research Center Incorporated Extracting candidate answers for a knowledge base from conversational sources
EP2953037A1 (en) * 2014-05-15 2015-12-09 Marco Palmonari Automated multi-grammar language processing system to optimize request handling in contact centers
US20160117311A1 (en) * 2014-10-22 2016-04-28 Thomson Licensing Method and Device for Performing Story Analysis
CN105706092A (en) * 2013-09-03 2016-06-22 罗杰·密德茂尔 Methods and systems of four-valued simulation
US9465793B2 (en) 2010-05-13 2016-10-11 Grammarly, Inc. Systems and methods for advanced grammar checking
US20160314506A1 (en) * 2015-04-23 2016-10-27 Korea University Research And Business Foundation Method, device, computer program and computer readable recording medium for determining opinion spam based on frame
US20170083508A1 (en) * 2015-09-18 2017-03-23 Mcafee, Inc. Systems and Methods for Multilingual Document Filtering
US20170083510A1 (en) * 2015-09-18 2017-03-23 Mcafee, Inc. Systems and Methods for Multi-Path Language Translation
US9665567B2 (en) 2015-09-21 2017-05-30 International Business Machines Corporation Suggesting emoji characters based on current contextual emotional state of user
US9667786B1 (en) * 2014-10-07 2017-05-30 Ipsoft, Inc. Distributed coordinated system and process which transforms data into useful information to help a user with resolving issues
US20170169094A1 (en) * 2015-12-15 2017-06-15 International Business Machines Corporation Statistical Clustering Inferred From Natural Language to Drive Relevant Analysis and Conversation With Users
US20170213247A1 (en) * 2012-08-28 2017-07-27 Nuance Communications, Inc. Systems and methods for engaging an audience in a conversational advertisement
US20170221483A1 (en) * 2010-05-13 2017-08-03 Alexander Poltorak Electronic personal interactive device
US20180011839A1 (en) * 2016-07-07 2018-01-11 Xerox Corporation Symbol prediction with gapped sequence models
US20180075131A1 (en) * 2016-09-13 2018-03-15 Microsoft Technology Licensing, Llc Computerized natural language query intent dispatching
US20180097940A1 (en) * 2016-09-30 2018-04-05 Genesys Telecommunications Laboratories, Inc. System and method for dynamic generation and optimization of process flows for a customer contact center
US20180189267A1 (en) * 2016-12-30 2018-07-05 Google Inc. Context-aware human-to-computer dialog
US10133733B2 (en) 2007-03-06 2018-11-20 Botanic Technologies, Inc. Systems and methods for an autonomous avatar driver
US10185754B2 (en) 2010-07-31 2019-01-22 Vocus Nm Llc Discerning human intent based on user-generated metadata
US20190042572A1 (en) * 2016-02-08 2019-02-07 Taiger Spain Sl System and method for querying questions and answers
US10249207B2 (en) 2016-01-19 2019-04-02 TheBeamer, LLC Educational teaching system and method utilizing interactive avatars with learning manager and authoring manager functions
US20190108290A1 (en) * 2017-10-06 2019-04-11 ClinMunications, LLC Human assisted automated question and answer system using natural language processing of real-time requests assisted by humans for requests of low confidence
US10282411B2 (en) * 2016-03-31 2019-05-07 International Business Machines Corporation System, method, and recording medium for natural language learning
US20190138594A1 (en) * 2017-11-06 2019-05-09 International Business Machines Corporation Pronoun Mapping for Sub-Context Rendering
US10421460B2 (en) * 2016-11-09 2019-09-24 Baidu Usa Llc Evaluation framework for decision making of autonomous driving vehicle
US10482490B2 (en) 2014-04-09 2019-11-19 Sailthru, Inc. Behavioral tracking system and method in support of high-engagement communications
US10503386B2 (en) 2008-05-21 2019-12-10 Please Don't Go, LLC. Messaging window overlay for a browser
US10540513B2 (en) 2016-09-13 2020-01-21 Microsoft Technology Licensing, Llc Natural language processor extension transmission data protection
CN111666444A (en) * 2020-06-02 2020-09-15 中国科学院计算技术研究所 Audio push method and system based on artificial intelligence, and related method and equipment
US10872091B2 (en) * 2018-11-16 2020-12-22 Fuvi Cognitive Network Corp. Apparatus, method, and system of cognitive data blocks and links for personalization, comprehension, retention, and recall of cognitive contents of a user
US10896294B2 (en) 2018-01-11 2021-01-19 End Cue, Llc Script writing and content generation tools and improved operation of same
US10922489B2 (en) * 2018-01-11 2021-02-16 RivetAI, Inc. Script writing and content generation tools and improved operation of same
US20210191926A1 (en) * 2019-12-20 2021-06-24 Visa International Service Association Query parsing from natural language questions supported by captured subject matter knowledge
US11068043B2 (en) 2017-07-21 2021-07-20 Pearson Education, Inc. Systems and methods for virtual reality-based grouping evaluation
US20210303772A1 (en) * 2020-03-31 2021-09-30 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and Apparatus for Constructing Document Heading Tree, Electronic Device and Storage Medium
US11314746B2 (en) 2013-03-15 2022-04-26 Cision Us Inc. Processing unstructured data streams using continuous queries
US20220222883A1 (en) * 2021-01-08 2022-07-14 Samsung Electronics Co., Ltd. Development platform for digital humans
US11442975B2 (en) * 2019-06-04 2022-09-13 Wipro Limited System and method for generating abstractive summary
US11707679B2 (en) * 2020-03-30 2023-07-25 Square Enix Co., Ltd. Medium, information processing apparatus, and method for generating a natural sentence
US11712628B2 (en) * 2018-09-20 2023-08-01 Apple Inc. Method and device for attenuation of co-user interactions

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9747636B2 (en) 2014-12-08 2017-08-29 Bank Of America Corporation Enhancing information security using an information passport dashboard
US10628521B2 (en) * 2015-08-03 2020-04-21 International Business Machines Corporation Scoring automatically generated language patterns for questions using synthetic events
US10628413B2 (en) * 2015-08-03 2020-04-21 International Business Machines Corporation Mapping questions to complex database lookups using synthetic events
US10528612B2 (en) * 2017-02-21 2020-01-07 International Business Machines Corporation Processing request documents
US11256754B2 (en) * 2019-12-09 2022-02-22 Salesforce.Com, Inc. Systems and methods for generating natural language processing training samples with inflectional perturbations
US11463657B1 (en) * 2020-11-10 2022-10-04 Know Systems Corp. System and method for an interactive digitally rendered avatar of a subject person

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030028498A1 (en) * 2001-06-07 2003-02-06 Barbara Hayes-Roth Customizable expert agent
US20030046689A1 (en) * 2000-09-25 2003-03-06 Maria Gaos Method and apparatus for delivering a virtual reality environment
US20060040718A1 (en) * 2004-07-15 2006-02-23 Mad Doc Software, Llc Audio-visual games and game computer programs embodying interactive speech recognition and methods related thereto
US20070288404A1 (en) * 2006-06-13 2007-12-13 Microsoft Corporation Dynamic interaction menus from natural language representations

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6341372B1 (en) * 1997-05-01 2002-01-22 William E. Datig Universal machine translator of arbitrary languages
US6233545B1 (en) * 1997-05-01 2001-05-15 William E. Datig Universal machine translator of arbitrary languages utilizing epistemic moments
US6778970B2 (en) * 1998-05-28 2004-08-17 Lawrence Au Topological methods to organize semantic network data flows for conversational applications
US6570555B1 (en) 1998-12-30 2003-05-27 Fuji Xerox Co., Ltd. Method and apparatus for embodied conversational characters with multimodal input/output in an interface device
US20020054072A1 (en) * 1999-12-15 2002-05-09 Barbara Hayes-Roth System, method, and device for an interactive messenger
WO2002037471A2 (en) * 2000-11-03 2002-05-10 Zoesis, Inc. Interactive character system
US7668718B2 (en) 2001-07-17 2010-02-23 Custom Speech Usa, Inc. Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile
US7610556B2 (en) * 2001-12-28 2009-10-27 Microsoft Corporation Dialog manager for interactive dialog with computer user
US7019749B2 (en) * 2001-12-28 2006-03-28 Microsoft Corporation Conversational interface agent
GB0201174D0 (en) * 2002-01-18 2002-03-06 British Telecomm Multi-mode interactive dialogue apparatus and method
US7076430B1 (en) * 2002-05-16 2006-07-11 At&T Corp. System and method of providing conversational visual prosody for talking heads
US7136818B1 (en) * 2002-05-16 2006-11-14 At&T Corp. System and method of providing conversational visual prosody for talking heads
US7113848B2 (en) * 2003-06-09 2006-09-26 Hanson David F Human emulation robot system
US20070061352A1 (en) * 2003-12-03 2007-03-15 Koninklijke Philips Electronic, N.V. System & method for integrative analysis of intrinsic and extrinsic audio-visual
US20060122834A1 (en) 2004-12-03 2006-06-08 Bennett Ian M Emotion detection device & method for use in distributed systems
US7827026B2 (en) * 2004-12-21 2010-11-02 Xerox Corporation Bilingual authoring assistant for the “tip of the tongue” problem
JP2006268375A (en) * 2005-03-23 2006-10-05 Fuji Xerox Co Ltd Translation memory system
WO2006105596A1 (en) * 2005-04-04 2006-10-12 Mor(F) Dynamics Pty Ltd Method for transforming language into a visual form
US8041570B2 (en) * 2005-05-31 2011-10-18 Robert Bosch Corporation Dialogue management using scripts
US20060282317A1 (en) * 2005-06-10 2006-12-14 Outland Research Methods and apparatus for conversational advertising
JP4087400B2 (en) * 2005-09-15 2008-05-21 株式会社東芝 Spoken dialogue translation apparatus, spoken dialogue translation method, and spoken dialogue translation program
JP4846336B2 (en) * 2005-10-21 2011-12-28 株式会社ユニバーサルエンターテインメント Conversation control device
US20080221892A1 (en) 2007-03-06 2008-09-11 Paco Xander Nathan Systems and methods for an autonomous avatar driver

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030046689A1 (en) * 2000-09-25 2003-03-06 Maria Gaos Method and apparatus for delivering a virtual reality environment
US20030028498A1 (en) * 2001-06-07 2003-02-06 Barbara Hayes-Roth Customizable expert agent
US20060040718A1 (en) * 2004-07-15 2006-02-23 Mad Doc Software, Llc Audio-visual games and game computer programs embodying interactive speech recognition and methods related thereto
US20070288404A1 (en) * 2006-06-13 2007-12-13 Microsoft Corporation Dynamic interaction menus from natural language representations

Cited By (140)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050086046A1 (en) * 1999-11-12 2005-04-21 Bennett Ian M. System & method for natural language processing of sentence based queries
US7624007B2 (en) * 1999-11-12 2009-11-24 Phoenix Solutions, Inc. System and method for natural language processing of sentence based queries
US10133733B2 (en) 2007-03-06 2018-11-20 Botanic Technologies, Inc. Systems and methods for an autonomous avatar driver
US20090030775A1 (en) * 2007-07-26 2009-01-29 Braintexter, Inc. System to generate and set up an advertising campaign based on the insertion of advertising messages within an exchange of messages, and method to operate said system
US8359234B2 (en) * 2007-07-26 2013-01-22 Braintexter, Inc. System to generate and set up an advertising campaign based on the insertion of advertising messages within an exchange of messages, and method to operate said system
US20130103501A1 (en) * 2007-07-26 2013-04-25 Braintexter, Inc. System to Generate and Set Up an Advertising Campaign Based on the Insertion of Advertising Messages within an Exchange of Messages, and Method to Operate Said System
US8909545B2 (en) * 2007-07-26 2014-12-09 Braintexter, Inc. System to generate and set up an advertising campaign based on the insertion of advertising messages within an exchange of messages, and method to operate said system
US20090037822A1 (en) * 2007-07-31 2009-02-05 Qurio Holdings, Inc. Context-aware shared content representations
US20090058862A1 (en) * 2007-08-27 2009-03-05 Finn Peter G Automatic avatar transformation for a virtual universe
US8261307B1 (en) 2007-10-25 2012-09-04 Qurio Holdings, Inc. Wireless multimedia content brokerage service for real time selective content provisioning
US8695044B1 (en) 2007-10-25 2014-04-08 Qurio Holdings, Inc. Wireless multimedia content brokerage service for real time selective content provisioning
US20090186635A1 (en) * 2008-01-22 2009-07-23 Braintexter, Inc. Systems and methods of contextual advertising
US20120179550A1 (en) * 2008-01-22 2012-07-12 Braintexter, Inc. Systems and methods of contextual advertising
US8423412B2 (en) * 2008-01-22 2013-04-16 Braintexter, Inc. Systems and methods of contextual advertising
US8156005B2 (en) * 2008-01-22 2012-04-10 Braintexter, Inc. Systems and methods of contextual advertising
US9110890B2 (en) 2008-02-15 2015-08-18 International Business Machines Corporation Selecting a language encoding of a static communication in a virtual universe
US20090210213A1 (en) * 2008-02-15 2009-08-20 International Business Machines Corporation Selecting a language encoding of a static communication in a virtual universe
US20090210803A1 (en) * 2008-02-15 2009-08-20 International Business Machines Corporation Automatically modifying communications in a virtual universe
US8156060B2 (en) * 2008-02-27 2012-04-10 Inteliwise Sp Z.O.O. Systems and methods for generating and implementing an interactive man-machine web interface based on natural language processing and avatar virtual agent based character
US20090216691A1 (en) * 2008-02-27 2009-08-27 Inteliwise Sp Z.O.O. Systems and Methods for Generating and Implementing an Interactive Man-Machine Web Interface Based on Natural Language Processing and Avatar Virtual Agent Based Character
US20090216625A1 (en) * 2008-02-27 2009-08-27 Adam Jeffrey Erlebacher Systems and Methods for Automated Identification and Evaluation of Brand Integration Opportunities in Scripted Entertainment
US20110213642A1 (en) * 2008-05-21 2011-09-01 The Delfin Project, Inc. Management system for a conversational system
US10503386B2 (en) 2008-05-21 2019-12-10 Please Don't Go, LLC. Messaging window overlay for a browser
US8738739B2 (en) 2008-05-21 2014-05-27 The Delfin Project, Inc. Automatic message selection with a chatbot
US8949377B2 (en) * 2008-05-21 2015-02-03 The Delfin Project, Inc. Management system for a conversational system
US20100058183A1 (en) * 2008-09-02 2010-03-04 International Business Machines Corporation Method, system, and program product for allocating virtual universe customer service
US8527430B2 (en) 2008-09-02 2013-09-03 International Business Machines Corporation Allocating virtual universe customer service
US8296246B2 (en) * 2008-09-02 2012-10-23 International Business Machines Corporation Allocating virtual universe customer service
US9195640B1 (en) 2009-01-12 2015-11-24 Sri International Method and system for finding content having a desired similarity
US8628392B1 (en) * 2009-05-22 2014-01-14 Xiaohui Kong Game of actual planning, task/time management, and information sharing
US9542389B2 (en) 2009-09-30 2017-01-10 International Business Machines Corporation Language translation in an environment associated with a virtual application
US8655644B2 (en) * 2009-09-30 2014-02-18 International Business Machines Corporation Language translation in an environment associated with a virtual application
US20110077934A1 (en) * 2009-09-30 2011-03-31 International Business Machines Corporation Language Translation in an Environment Associated with a Virtual Application
US10303774B2 (en) 2009-10-29 2019-05-28 Unify Gmbh & Co. Kg Method and system to automatically change or update the configuration or setting of a communication system
KR20170008317A (en) * 2009-10-29 2017-01-23 유니파이 게엠베하 운트 코. 카게 Method and system to automatically change or update the configuration or setting of a communication system
KR102009912B1 (en) 2009-10-29 2019-08-12 유니파이 게엠베하 운트 코. 카게 Method and system to automatically change or update the configuration or setting of a communication system
CN102055793A (en) * 2009-10-29 2011-05-11 西门子企业通讯有限责任两合公司 Method and system to automatically change or update the configuration or setting of a communication system
US20110106528A1 (en) * 2009-10-29 2011-05-05 Siemens Enterprise Communications Gmbh & Co.Kg Method and System to Automatically Change or Update the Configuration or Setting of a Communication System
EP2317730A1 (en) * 2009-10-29 2011-05-04 Siemens Enterprise Communications GmbH & Co. KG Method and system to automatically change or update the configuration or setting of a communication system
US10650194B2 (en) 2009-10-29 2020-05-12 Unify Gmbh & Co. Kg Method and system to automatically change or update the configuration or setting of a communication system
US20110137844A1 (en) * 2009-12-07 2011-06-09 Alphaport, Inc. Avatar-Based Technical Networking System
US8560482B2 (en) 2009-12-07 2013-10-15 Alphaport, Inc. Avatar-based technical networking system
US20110258544A1 (en) * 2010-04-16 2011-10-20 Avaya Inc. System and method for suggesting automated assistants based on a similarity vector in a graphical user interface for managing communication sessions
US10079892B2 (en) * 2010-04-16 2018-09-18 Avaya Inc. System and method for suggesting automated assistants based on a similarity vector in a graphical user interface for managing communication sessions
US11367435B2 (en) * 2010-05-13 2022-06-21 Poltorak Technologies Llc Electronic personal interactive device
US20170221484A1 (en) * 2010-05-13 2017-08-03 Alexander Poltorak Electronic personal interactive device
US10387565B2 (en) 2010-05-13 2019-08-20 Grammarly, Inc. Systems and methods for advanced grammar checking
US9465793B2 (en) 2010-05-13 2016-10-11 Grammarly, Inc. Systems and methods for advanced grammar checking
US20170221483A1 (en) * 2010-05-13 2017-08-03 Alexander Poltorak Electronic personal interactive device
US11341962B2 (en) * 2010-05-13 2022-05-24 Poltorak Technologies Llc Electronic personal interactive device
US10185754B2 (en) 2010-07-31 2019-01-22 Vocus Nm Llc Discerning human intent based on user-generated metadata
US8620850B2 (en) 2010-09-07 2013-12-31 Blackberry Limited Dynamically manipulating an emoticon or avatar
US20130218678A1 (en) * 2010-10-19 2013-08-22 Citizennet Inc Systems and methods for selecting and generating targeting information for specific advertisements based upon affinity information obtained from an online social network
US20120129612A1 (en) * 2010-11-19 2012-05-24 KABUSHIKI KAISHA SQUARE ENIX (also trading as "SQUARE ENIX CO., LTD.") Game apparatus, game program and information recording medium
US8317625B2 (en) * 2010-11-19 2012-11-27 Kabushiki Kaisha Square Enix Game apparatus, game program and information recording medium
US10162891B2 (en) * 2010-11-29 2018-12-25 Vocus Nm Llc Determining demographics based on user interaction
US20140114998A1 (en) * 2010-11-29 2014-04-24 Viralheat, Inc. Determining demographics based on user interaction
US9189565B2 (en) 2010-11-30 2015-11-17 International Business Machines Corporation Managing tag clouds
US8769037B2 (en) 2010-11-30 2014-07-01 International Business Machines Corporation Managing tag clouds
US9063927B2 (en) 2011-04-06 2015-06-23 Citizennet Inc. Short message age classification
US9836447B2 (en) * 2011-07-28 2017-12-05 Microsoft Technology Licensing, Llc Linguistic error detection
US20150006159A1 (en) * 2011-07-28 2015-01-01 Microsoft Corporation Linguistic error detection
US9594850B2 (en) 2011-11-17 2017-03-14 Sri International Method and system utilizing a personalized user model to develop a search request
US8868590B1 (en) 2011-11-17 2014-10-21 Sri International Method and system utilizing a personalized user model to develop a search request
US9135245B1 (en) * 2012-01-09 2015-09-15 Cox Communications, Inc. Filtering content based on acquiring data associated with language identification
US20160328378A1 (en) * 2012-08-02 2016-11-10 American Express Travel Related Services Company, Inc. Anaphora resolution for semantic tagging
US9280520B2 (en) * 2012-08-02 2016-03-08 American Express Travel Related Services Company, Inc. Systems and methods for semantic information retrieval
US9424250B2 (en) * 2012-08-02 2016-08-23 American Express Travel Related Services Company, Inc. Systems and methods for semantic information retrieval
US9805024B2 (en) * 2012-08-02 2017-10-31 American Express Travel Related Services Company, Inc. Anaphora resolution for semantic tagging
US20140039877A1 (en) * 2012-08-02 2014-02-06 American Express Travel Related Services Company, Inc. Systems and Methods for Semantic Information Retrieval
US20160132483A1 (en) * 2012-08-02 2016-05-12 American Express Travel Related Services Company, Inc. Systems and methods for semantic information retrieval
US20170213247A1 (en) * 2012-08-28 2017-07-27 Nuance Communications, Inc. Systems and methods for engaging an audience in a conversational advertisement
US20140278605A1 (en) * 2013-03-15 2014-09-18 Ncr Corporation System and method of completing an activity via an agent
US11314746B2 (en) 2013-03-15 2022-04-26 Cision Us Inc. Processing unstructured data streams using continuous queries
US10726461B2 (en) * 2013-03-15 2020-07-28 Ncr Corporation System and method of completing an activity via an agent
US20150004591A1 (en) * 2013-06-27 2015-01-01 DoSomething.Org Device, system, method, and computer-readable medium for providing an educational, text-based interactive game
US20150051910A1 (en) * 2013-08-19 2015-02-19 Nuance Communications, Inc. Unsupervised Clustering of Dialogs Extracted from Released Application Logs
US9606984B2 (en) * 2013-08-19 2017-03-28 Nuance Communications, Inc. Unsupervised clustering of dialogs extracted from released application logs
US9865255B2 (en) * 2013-08-29 2018-01-09 Panasonic Intellectual Property Corporation Of America Speech recognition method and speech recognition apparatus
US10446151B2 (en) 2013-08-29 2019-10-15 Panasonic Intellectual Property Corporation Of America Speech recognition method and speech recognition apparatus
US20150262577A1 (en) * 2013-08-29 2015-09-17 Panasonic Intellectual Property Corporation Of America Speech recognition method and speech recognition apparatus
US9576319B2 (en) * 2013-09-03 2017-02-21 Roger Midmore Methods and systems of four-valued Monte Carlo simulation for financial modeling
US20150066730A1 (en) * 2013-09-03 2015-03-05 Roger Midmore Methods and Systems of Four-Valued Monte Carlo Simulation for Financial Modeling
US20150066476A1 (en) * 2013-09-03 2015-03-05 Roger Midmore Methods and Systems of Four Valued Analogical Transformation Operators Used in Natural Language Processing and Other Applications
US9672466B2 (en) * 2013-09-03 2017-06-06 Roger Midmore Methods and systems of four-valued genomic sequencing and macromolecular analysis
CN105706091A (en) * 2013-09-03 2016-06-22 罗杰·密德茂尔 Methods and systems of four valued analogical transformation operators used in natural language processing and other applications
US20150066835A1 (en) * 2013-09-03 2015-03-05 Roger Midmore Computational Tools For Genomic Sequencing and Macromolecular Analysis
CN105706092A (en) * 2013-09-03 2016-06-22 罗杰·密德茂尔 Methods and systems of four-valued simulation
US9575951B2 (en) * 2013-09-03 2017-02-21 Roger Midmore Methods and systems of four valued analogical transformation operators used in natural language processing and other applications
WO2015039214A1 (en) * 2013-09-18 2015-03-26 Moldoveanu Mihnea Calin System and method for providing a communication platform with guided dialogical functions
US20150100381A1 (en) * 2013-10-03 2015-04-09 Douglas Petrie Method and System for Increasing the Percentage of Customers Realized from Smart Phone Advertising
US20150193550A1 (en) * 2014-01-08 2015-07-09 International Business Machines Corporation Presenting tags of a tag cloud in a more understandable and visually appealing manner
US9910917B2 (en) * 2014-01-08 2018-03-06 International Business Machines Corporation Presenting tags of a tag cloud in a more understandable and visually appealing manner
US9910918B2 (en) * 2014-01-08 2018-03-06 International Business Machines Corporation Presenting tags of a tag cloud in a more understandable and visually appealing manner
US20150193415A1 (en) * 2014-01-08 2015-07-09 International Business Machines Corporation Presenting tags of a tag cloud in a more understandable and visually appealing manner
US20150248397A1 (en) * 2014-02-28 2015-09-03 Educational Testing Service Computer-Implemented Systems and Methods for Measuring Discourse Coherence
US9665566B2 (en) * 2014-02-28 2017-05-30 Educational Testing Service Computer-implemented systems and methods for measuring discourse coherence
US10482490B2 (en) 2014-04-09 2019-11-19 Sailthru, Inc. Behavioral tracking system and method in support of high-engagement communications
EP2953037A1 (en) * 2014-05-15 2015-12-09 Marco Palmonari Automated multi-grammar language processing system to optimize request handling in contact centers
US20150340026A1 (en) * 2014-05-22 2015-11-26 Palo Alto Research Center Incorporated Extracting candidate answers for a knowledge base from conversational sources
US9667786B1 (en) * 2014-10-07 2017-05-30 Ipsoft, Inc. Distributed coordinated system and process which transforms data into useful information to help a user with resolving issues
US20160117311A1 (en) * 2014-10-22 2016-04-28 Thomson Licensing Method and Device for Performing Story Analysis
US20160314506A1 (en) * 2015-04-23 2016-10-27 Korea University Research And Business Foundation Method, device, computer program and computer readable recording medium for determining opinion spam based on frame
US9928236B2 (en) * 2015-09-18 2018-03-27 Mcafee, Llc Systems and methods for multi-path language translation
US9984068B2 (en) * 2015-09-18 2018-05-29 Mcafee, Llc Systems and methods for multilingual document filtering
US20170083508A1 (en) * 2015-09-18 2017-03-23 Mcafee, Inc. Systems and Methods for Multilingual Document Filtering
US20170083510A1 (en) * 2015-09-18 2017-03-23 Mcafee, Inc. Systems and Methods for Multi-Path Language Translation
US9665567B2 (en) 2015-09-21 2017-05-30 International Business Machines Corporation Suggesting emoji characters based on current contextual emotional state of user
US20170169094A1 (en) * 2015-12-15 2017-06-15 International Business Machines Corporation Statistical Clustering Inferred From Natural Language to Drive Relevant Analysis and Conversation With Users
US9940384B2 (en) * 2015-12-15 2018-04-10 International Business Machines Corporation Statistical clustering inferred from natural language to drive relevant analysis and conversation with users
US10249207B2 (en) 2016-01-19 2019-04-02 TheBeamer, LLC Educational teaching system and method utilizing interactive avatars with learning manager and authoring manager functions
US20190042572A1 (en) * 2016-02-08 2019-02-07 Taiger Spain Sl System and method for querying questions and answers
US10592504B2 (en) * 2016-02-08 2020-03-17 Capricorn Holdings Pte, Ltd. System and method for querying questions and answers
US10282411B2 (en) * 2016-03-31 2019-05-07 International Business Machines Corporation System, method, and recording medium for natural language learning
US20180011839A1 (en) * 2016-07-07 2018-01-11 Xerox Corporation Symbol prediction with gapped sequence models
US20180075131A1 (en) * 2016-09-13 2018-03-15 Microsoft Technology Licensing, Llc Computerized natural language query intent dispatching
US10503767B2 (en) * 2016-09-13 2019-12-10 Microsoft Technology Licensing, Llc Computerized natural language query intent dispatching
US10540513B2 (en) 2016-09-13 2020-01-21 Microsoft Technology Licensing, Llc Natural language processor extension transmission data protection
US20180097940A1 (en) * 2016-09-30 2018-04-05 Genesys Telecommunications Laboratories, Inc. System and method for dynamic generation and optimization of process flows for a customer contact center
US10421460B2 (en) * 2016-11-09 2019-09-24 Baidu Usa Llc Evaluation framework for decision making of autonomous driving vehicle
US10268680B2 (en) * 2016-12-30 2019-04-23 Google Llc Context-aware human-to-computer dialog
US20180189267A1 (en) * 2016-12-30 2018-07-05 Google Inc. Context-aware human-to-computer dialog
US11227124B2 (en) 2016-12-30 2022-01-18 Google Llc Context-aware human-to-computer dialog
US11068043B2 (en) 2017-07-21 2021-07-20 Pearson Education, Inc. Systems and methods for virtual reality-based grouping evaluation
US20190108290A1 (en) * 2017-10-06 2019-04-11 ClinMunications, LLC Human assisted automated question and answer system using natural language processing of real-time requests assisted by humans for requests of low confidence
US20190138594A1 (en) * 2017-11-06 2019-05-09 International Business Machines Corporation Pronoun Mapping for Sub-Context Rendering
US10546060B2 (en) * 2017-11-06 2020-01-28 International Business Machines Corporation Pronoun mapping for sub-context rendering
US10671808B2 (en) * 2017-11-06 2020-06-02 International Business Machines Corporation Pronoun mapping for sub-context rendering
US10896294B2 (en) 2018-01-11 2021-01-19 End Cue, Llc Script writing and content generation tools and improved operation of same
US10922489B2 (en) * 2018-01-11 2021-02-16 RivetAI, Inc. Script writing and content generation tools and improved operation of same
US11712628B2 (en) * 2018-09-20 2023-08-01 Apple Inc. Method and device for attenuation of co-user interactions
US10872091B2 (en) * 2018-11-16 2020-12-22 Fuvi Cognitive Network Corp. Apparatus, method, and system of cognitive data blocks and links for personalization, comprehension, retention, and recall of cognitive contents of a user
US11442975B2 (en) * 2019-06-04 2022-09-13 Wipro Limited System and method for generating abstractive summary
US20210191926A1 (en) * 2019-12-20 2021-06-24 Visa International Service Association Query parsing from natural language questions supported by captured subject matter knowledge
US11586618B2 (en) * 2019-12-20 2023-02-21 Visa International Service Association Query parsing from natural language questions supported by captured subject matter knowledge
US11707679B2 (en) * 2020-03-30 2023-07-25 Square Enix Co., Ltd. Medium, information processing apparatus, and method for generating a natural sentence
US20210303772A1 (en) * 2020-03-31 2021-09-30 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and Apparatus for Constructing Document Heading Tree, Electronic Device and Storage Medium
CN111666444A (en) * 2020-06-02 2020-09-15 中国科学院计算技术研究所 Audio push method and system based on artificial intelligence, and related method and equipment
US20220222883A1 (en) * 2021-01-08 2022-07-14 Samsung Electronics Co., Ltd. Development platform for digital humans
US11893669B2 (en) * 2021-01-08 2024-02-06 Samsung Electronics Co., Ltd. Development platform for digital humans

Also Published As

Publication number Publication date
US20150066484A1 (en) 2015-03-05
US10133733B2 (en) 2018-11-20

Similar Documents

Publication Publication Date Title
US10133733B2 (en) Systems and methods for an autonomous avatar driver
Rajput Natural language processing, sentiment analysis, and clinical analytics
Ignatow et al. Text mining: A guidebook for the social sciences
CN109314660B (en) Method and device for providing news recommendation in automatic chat
US20170017635A1 (en) Natural language processing system and method
US10496756B2 (en) Sentence creation system
Millstein Natural language processing with python: natural language processing using NLTK
US10013404B2 (en) Targeted story summarization using natural language processing
US20140046876A1 (en) System and method of providing a computer-generated response
Syed et al. Associating targets with SentiUnits: a step forward in sentiment analysis of Urdu text
Bellot et al. INEX Tweet Contextualization task: Evaluation, results and lesson learned
Krommyda et al. Emotion detection in Twitter posts: a rule-based algorithm for annotated data acquisition
Trye et al. Harnessing Indigenous Tweets: The Reo Māori Twitter corpus
Da et al. Deep learning based dual encoder retrieval model for citation recommendation
Demir et al. Graph-based Turkish text normalization and its impact on noisy text processing
Ho et al. Extracting lexical and phrasal paraphrases: a review of the literature
Kumar et al. Pattern-based syntactic simplification of compound and complex sentences
CN114255067A (en) Data pricing method and device, electronic equipment and storage medium
Ennaji et al. A product reputation framework based on social multimedia content
Jenkins Designing Service-Oriented Chatbot Systems Using a Construction Grammar-Driven Natural Language Generation System
Mohamed et al. SDbQfSum: Query‐focused summarization framework based on diversity and text semantic analysis
Bansal et al. Online Insurance Business Analytics Approach for Customer Segmentation
Tanantong et al. A Survey of Automatic Text Classification Based on Thai Social Media Data
Swanson Enabling open domain interactive storytelling using a data-driven case-based approach
US20230289377A1 (en) Multi-channel feedback analytics for presentation generation

Legal Events

Date Code Title Description
AS Assignment

Owner name: HEADCASE HUMANUFACTURING, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NATHAN, PACO XANDER;COX, CATHI JOANN;LEIBERT, FLORIAN THOMAS;AND OTHERS;REEL/FRAME:020662/0910;SIGNING DATES FROM 20080215 TO 20080225

AS Assignment

Owner name: OGG, LLC, DELAWARE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HEADCASE HUMANUFACTURING, INC.;REEL/FRAME:033782/0705

Effective date: 20140918

AS Assignment

Owner name: MEADOWS, MARK STEPHEN, DELAWARE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:OGG, LLC;REEL/FRAME:033791/0104

Effective date: 20140709

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: BOTANIC TECHNOLOGIES, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MEADOWS, MARK STEPHEN;REEL/FRAME:042162/0456

Effective date: 20170425

AS Assignment

Owner name: BOTANIC TECHNOLOGIES, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HEADCASE HUMANUFACTURING, INC.;REEL/FRAME:042800/0809

Effective date: 20170620