US20130179439A1 - Methods and Systems for Utilizing Contextual Feedback to Generate and Modify Playlists - Google Patents

Methods and Systems for Utilizing Contextual Feedback to Generate and Modify Playlists Download PDF

Info

Publication number
US20130179439A1
US20130179439A1 US13/669,412 US201213669412A US2013179439A1 US 20130179439 A1 US20130179439 A1 US 20130179439A1 US 201213669412 A US201213669412 A US 201213669412A US 2013179439 A1 US2013179439 A1 US 2013179439A1
Authority
US
United States
Prior art keywords
vector
song
database
item
songs
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US13/669,412
Inventor
Etienne F. Handman
Thomas J. Conrad
Joseph J. Kennedy
Timothy B. Westergren
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Pandora Media LLC
Original Assignee
Pandora Media LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US10/150,876 external-priority patent/US7003515B1/en
Priority claimed from US11/295,339 external-priority patent/US20060206478A1/en
Priority claimed from US11/279,567 external-priority patent/US20060212442A1/en
Application filed by Pandora Media LLC filed Critical Pandora Media LLC
Priority to US13/669,412 priority Critical patent/US20130179439A1/en
Publication of US20130179439A1 publication Critical patent/US20130179439A1/en
Assigned to WELLS FARGO BANK, NATIONAL ASSOCIATION reassignment WELLS FARGO BANK, NATIONAL ASSOCIATION SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PANDORA MEDIA CALIFORNIA, LLC, PANDORA MEDIA, INC.
Assigned to PANDORA MEDIA, INC. reassignment PANDORA MEDIA, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CONRAD, THOMAS J., HANDMAN, ETIENNE F., KENNEDY, JOSEPH J., WESTERGREN, TIMOTHY B.
Assigned to PANDORA MEDIA CALIFORNIA, LLC, ADSWIZ INC. reassignment PANDORA MEDIA CALIFORNIA, LLC RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: WELLS FARGO BANK, NATIONAL ASSOCIATION
Assigned to PANDORA MEDIA, LLC reassignment PANDORA MEDIA, LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: PANDORA MEDIA, INC.
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06F17/30554
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/248Presentation of query results
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/632Query formulation
    • G06F16/634Query by example, e.g. query by humming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/638Presentation of query results
    • G06F16/639Presentation of query results using playlists
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/64Browsing; Visualisation therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/686Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title or artist information, time, location or usage information, user ratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/687Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/034Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • G11B27/105Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/11Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier

Definitions

  • Embodiments of the invention are directed to methods and systems for utilizing contextual feedback to generate and modify playlists.
  • Embodiments of the invention are directed to methods and systems for using contextual information to generate and modify playlists.
  • one embodiment of the invention includes the steps of storing feedback from at least one user about a first content object; storing contextual information relating to the feedback; retrieving the feedback and contextual information; and generating a playlist of content objects using the feedback and contextual information.
  • Another embodiment of the invention includes the steps of comparing an input seed to a first database item, wherein the first database item corresponds to a first content object; retrieving contextual information relating to the first content object; and selecting the first content object for the playlist based on the comparison and the contextual information.
  • computer-executable instructions for implementing the disclosed methods are stored as control logic or computer-readable instructions on computer-readable media, such as an optical or magnetic disk.
  • FIG. 1 depicts an exemplary operating environment for an embodiment of the invention
  • FIGS. 2 a and 2 b depict terminal-based displays for presenting and providing content to a user in accordance with embodiments of the invention
  • FIGS. 3 a - 3 d depict in more detail the graphical user interface of FIGS. 2 a and 2 b in various stages of operation and in accordance with an embodiment of the invention
  • FIG. 4 depicts, in accordance with an embodiment of the invention, a station pop-up menu generated in response to a user selecting a button such as “Station 1 ” button 308 in FIG. 3 c.
  • FIG. 5 depicts, in accordance with an embodiment of the invention, the graphical user interface of FIGS. 2 a and 2 b after a user has clicked the “Add More Music” menu choice 402 of station pop-up menu 400 in FIG. 4 ;
  • FIG. 6 depicts, in accordance with an embodiment of the invention, the graphical user interface of FIGS. 2 a and 2 b after a user has clicked the “Email This Station” menu choice 404 of station pop-up menu 400 in FIG. 4 ;
  • FIGS. 7 a - c depict, in accordance with an embodiment of the invention, the graphical user interface of FIGS. 2 a and 2 b in various stages of operation after a user has clicked the “Edit This Station” menu choice 406 of station pop-up menu 400 in FIG. 4 ;
  • FIG. 8 depicts, in accordance with an embodiment of the invention, a content pop-up menu generated in response to a user selecting a component of a graphical element, such as content art 332 of second graphic element 326 .
  • FIG. 9 depicts, in accordance with an embodiment of the invention, an “Information” panel 900 that appears on graphical user interface 208 after the user has selected, for example, “Why Did You Play This Song” menu choice 802 .
  • FIG. 10 depicts, in accordance with an embodiment of the invention, a “Create New Station” panel 1000 that appears on graphical user interface 208 after the user has selected, for example, “Make a New Station from This Song” menu choice 804 .
  • FIG. 11 depicts a “Favorites” display 1100 in accordance with an embodiment of the invention.
  • FIG. 12 depicts a flow diagram overview of methods for presenting and providing content to a user.
  • FIG. 13 depicts a relationship between different song candidates.
  • FIG. 14 is a graph showing a deviation vector.
  • FIG. 15 graphically depicts a bimodal song group.
  • FIG. 16 shows a flow diagram for one or more embodiments of the “Generate or Modify Playlist” step 1204 in FIG. 12 .
  • FIG. 17 shows a flow diagram for one or more embodiments of the “Identify Characteristics” step 1604 in FIG. 16 .
  • FIG. 18 shows a flow diagram of a method of generating a playlist of content objects in accordance with aspects of the invention.
  • FIG. 19 depicts feedback and contextual information stored in accordance with aspects of the invention.
  • FIG. 20 shows another flow diagram of a method of generating a playlist of content objects in accordance with aspects of the invention.
  • FIG. 1 depicts a diagram of exemplary system 100 that may be used to implement embodiments of the invention.
  • a plurality of terminals such as terminals 102 , 104 and 106 , coupled to playlist server 108 and content server 118 via network 110 .
  • playlist server 108 and content server 118 may be the same server performing all functions of playlist server 108 and content server 118 .
  • Terminals 102 , 104 and 106 , playlist server 108 and content server 118 may include a processor, memory and other conventional electronic components and may be programmed with processor-executable instructions to facilitate communication via network 110 and perform aspects of the invention.
  • network 110 is not limited to a particular type of network.
  • network 110 may feature one or more wide area networks (WANs), such as the Internet.
  • WANs wide area networks
  • Network 110 may also feature one or more local area networks (LANs) having one or more of the well-known LAN topologies and the use of a variety of different protocols on these topologies, such as Ethernet, TCP/IP, Frame Relay, Ethernet, FTP, HTTP and the like, is presumed.
  • LANs local area networks
  • network 110 may feature a Public Switched Telephone Network (PSTN) featuring land-line and cellular telephone terminals, or else a network featuring a combination of any or all of the above.
  • PSTN Public Switched Telephone Network
  • Terminals 102 , 104 and 106 may be coupled to network 110 via, for example, twisted pair wires, coaxial cable, fiber optics, electromagnetic waves or other media.
  • playlist server 108 contains a database of items 112 .
  • playlist server 108 may be coupled to database of items 112 .
  • playlist server 108 may be coupled to a “MUSIC GENOME PROJECT” database as described in U.S. Pat. No. 7,003,515.
  • Playlist server 108 may also contain or be coupled to matching engine 114 .
  • Matching engine 114 utilizes an associated set of search and matching functions 116 to operate on the database of items 112 .
  • matching engine 114 utilizes search and matching functions implemented in software or hardware to effectively calculate the distance between a source song and other songs in the database (as described here and in U.S. Pat. No. 7,003,515), and then sorts the results to yield an adjustable number of closest matches.
  • content server 118 contains a database of content objects 120 .
  • content server 118 may be wholly or partially integrated with playlist server 108 , or separately coupled to a database of content objects 120 .
  • Content server 118 may also contain or be coupled to content engine 122 .
  • Content engine 122 utilizes an associated set of management functions 124 , such as standard finding, packaging and sending functions, to operate on the database of content objects 122 .
  • management functions 124 such as standard finding, packaging and sending functions
  • Terminals 102 , 104 and 106 feature user interfaces that enable users to interact with server 108 .
  • the user interfaces may allow users to utilize a variety of functions, such as displaying information from playlist server 108 , requesting additional information from playlist server 108 , customizing local and/or remote aspects of the system and controlling local and/or remote aspects of the system.
  • Terminals 102 , 104 and 106 can be operated in a client-server configuration to permit a user to retrieve web pages from playlist server 108 .
  • any of various conventional web browsers can be used to display and manipulate data on the web pages.
  • FIG. 2 a depicts terminal-based display 200 for presenting and providing content to a user in accordance with an embodiment of the invention.
  • Terminal-based display 200 may comprise, for example, a web browser window 204 displayed on terminal 102 ( FIG. 1 ) running an operating system such as “WINDOWS” from Microsoft Corp.
  • terminal 102 is configured as the client in a client/server relationship with playlist server 108 and content server 118 .
  • a user of terminal 102 establishes a client/server relationship with playlist server 108 by inputting the appropriate URL in address field 206 (in this case, the URL is “http://www.pandora.com”).
  • the URL is “http://www.pandora.com”.
  • web page 204 is retrieved from playlist server 108 .
  • web page 204 features graphical user interface 208 (shown in more detail in, e.g., FIG. 3 d ), “favorites” button 210 , “minimize” button 212 , tip 214 and advertisement 216 .
  • terminal-based display 220 presents and provides content to a user in accordance with another embodiment of the invention.
  • terminal-based display 220 may comprise, for example, a web browser window 222 featuring graphical user interface 208 without, for example, “favorites” button 210 , “minimize” button 212 , tip 214 and advertisement 216 .
  • Terminal-based display 220 is smaller than terminal-based display 200 and thus better preserves desktop display resources.
  • the user is given the option to return graphical user interface 208 to terminal-based display 200 .
  • the user of terminal 102 may discontinue the client/server relationship with playlist server 108 by selecting “close window” button 218 .
  • playlist server 108 recognizes the user as a result of well-known schemes such as “cookies” and thus retains any customized user preferences or settings when web page 204 is retrieved and graphical user interface 208 is restarted.
  • tip 214 enhances the user-friendliness of graphical user interface 208 by providing information to the user regarding how to use graphical user interface 208 .
  • tip 214 may state “Use thumbs up/thumbs down to tune your stations. Click here to learn more.” To the extent the user clicks the hypertext link “Click here,” another web page is retrieved providing more detailed information about how to tune stations.
  • Tip 214 may also advertise career opportunities or display other information.
  • tip 214 may be provided in connection with terminal-based display 220 .
  • advertisement 216 may comprise a standard paid “banner” advertisement for a third party in any configuration on web page 204 .
  • Advertisement 216 may generate royalty revenue or other income for the operator.
  • the type of advertisement 216 presented to the user on web page 204 depends on various criteria, including but not limited to input, feedback and other information provided by the user, the location of the user's IP address, and other information such as the time of day or year.
  • FIGS. 3 a - 3 d depict in more detail graphical user interface 208 ( FIGS. 2 a and 2 b ) in various stages of operation and in accordance with an embodiment of the invention.
  • Graphical user interface 208 is provided through playlist server 108 ( FIG. 1 ) and may be implemented through, for example, Java, JavaScript, XML, Flash or HTML.
  • Playlist panel 304 features “Add to Playlist” button 306 and “Playlist 1 ,” “Playlist 2 ” and “Playlist 3 ” buttons 308 , 310 and 312 .
  • “Create Station” button 306 initiates the generation of a station (e.g., a station corresponding to “Station 1 ” button 308 ) corresponding to an input seed, such as a song name or artist name, selectively provided by the user.
  • the station facilitates the providing of content to the user that, for example, corresponds to a playlist generated as a result of a comparison of the input seed to musicological attributes of other songs.
  • the user could input “Miles Davis” and a “Miles Davis station” would be created that facilitates the providing of content to the user that corresponds to “Miles Davis” songs or songs that are musicologically similar to songs by “Miles Davis.”
  • playlist panel 304 visually represents to the user a playlist of content objects such as songs, the first song of which corresponds to first graphic element 314 and the second song of which corresponds to second graphic element 326 .
  • First graphic element features corresponding song text 316 , artist text 318 and content art 320
  • second graphic element 326 features corresponding song text 328 , artist text 330 and content art 332 .
  • Corresponding song text 316 and 328 , as well as corresponding artist text 318 and 330 may additionally comprise hypertext links that provide additional information, such as background knowledge about an artist or song.
  • Corresponding content art 320 and 332 may comprise, for example, a picture of an album cover.
  • first graphic element 314 or second graphic element 326 may feature additional or fewer components than the embodiment that has been described.
  • Other types of components include “purchase” buttons, advertisements, feedback indicators (such as feedback indicator 336 in FIG. 3 d ) and links to additional services and information.
  • other embodiments of first graphic element 314 or second graphic element 326 may feature different sizes, shapes and appearances than the embodiment that has been described.
  • the song currently being provided to the user is visually represented by the rightmost graphic element (i.e., second graphic element 326 ).
  • the graphic elements corresponding to those songs are scrolled to the left across playlist panel 304 (in this example, approximately three graphic elements total can be visualized to the user).
  • first graphic element 314 corresponds to a song that has already been provided to the user
  • second graphic element 326 corresponds to a song that is currently being provided to the user.
  • second graphic element 326 is currently being provided to the user is emphasized by tinting, shading or otherwise de-emphasizing first graphic element 314 , or highlighting, brightening or otherwise emphasizing second graphic element 326 .
  • playback bar 334 may be featured as a component of second graphic element 326 to indicate how much of the currently provided song has already been played.
  • other embodiments may feature alternative ways of visually representing the playlist and/or the progression of the playlist, as well as fewer or more graphic elements and alternative ways for representing those graphic elements.
  • volume control 340 adjusts the audible volume of content objects having audio that are provided to the user in accordance with embodiments of the invention.
  • playback controls 342 allow the user to pause or resume the playing of content objects. Playback controls 342 also allow the user to terminate playing of the current content object in favor of another content object.
  • the user's selecting of “Help” button 344 generates an on-screen pop-up menu providing clickable menu choices that provide additional features to the user and enhance the user-friendliness of graphical user interface 208 .
  • the on-screen pop-up menu may include choices providing additional information about a music discovery service, such as a FAQ, contact information or legal notices.
  • the user's selecting of “Share” button 346 generates another pop-up menu providing clickable menu choices relating to, for example, sharing features of graphical music interface 208 .
  • the pop-up menu may include choices for providing a playlist to other users of the music discovery service (e.g., enabling another user to enjoy a station such as the station corresponding to “Station 1 ” button 308 and thus to be provided content corresponding to that station).
  • the pop-up menu may also include choices for facilitating the providing of content by another station created by another user, the operator or a third party.
  • the user's selecting of “Account” button 348 generates another pop-up menu providing clickable menu choices relating to, for example, customized user preferences or settings.
  • the pop-up menu may include choices for viewing favorite stations, editing account and contact information or subscribing to the music discovery service. “Subscribing” may mean, for example, that in exchange for an annual fee, the user will no longer see advertisement 216 when using the music discovery service.
  • the user's selecting of “Guide” button 350 generates another pop-up menu providing clickable menu choices relating to, for example, enabling the user to selectively provide feedback about a content object such as a song.
  • “Guide” button 350 serves as the primary interface for the “back-and-forth” conversation between the user and the music discovery service.
  • the pop-up menu may include choices for enabling the user to provide feedback corresponding to comments such as “I really like this song,” “I don't like this song,” or “I′m tired of this song.” This feedback can be used to customize, adapt and/or enhance the initial playlist generated in connection with a station so that it is more attuned to the preferences of the user.
  • the pop-up menu generated by selecting “Guide” button 350 may include other feedback options, such as reasons why the user likes or dislikes a certain song.
  • reasons why the user likes or dislikes a certain song include “I like the artist,” “I like the song,” “I like the beat,” “I like the instrument being played,” “I like the meaning of the lyrics,” or “I like the genre.”
  • reasons why the user dislikes the song include “I don't like the artist,” “I don't like the vocals,” “I don't like the repetitiveness,” “The music is too ‘mainstream,’” or “The music is too loud.”
  • the playlist may be modified. Modifications to the playlist are accomplished, for example, by the use of weighing values and scaling functions as described in currently pending U.S. patent application Ser. No. 11/295,339, as will be discussed further below.
  • “Guide” button may include other choices that provide the user with information as to why a song is being played (i.e. what musicological attributes, such as characteristics or focus traits, are contained in a song). “Guide” button may also include other choices that enable the user to selectively modify the input seed so that it, and the playlist that is generated as a result of a comparison between the input seed and other songs, reflects additional artists or songs.
  • FIG. 4 depicts, in accordance with an embodiment of the invention, station pop-up menu 400 , which is generated in response to a user selecting a button such as “Station 1 ” button 308 in FIG. 3 c .
  • Station pop-up menu 400 includes menu choices such as “Add More Music” menu choice 402 , “Email This Station” menu choice 404 , “Edit This Station” menu choice 406 , “Rename This Station” menu choice 408 and “Delete This Station” menu choice 410 .
  • Other embodiments of the invention may have fewer, additional or alternative menu choices.
  • “Add More Music” menu choice 402 enables the user to selectively modify the input seed corresponding to the current station.
  • FIG. 5 depicts “Add More Music” panel 500 that appears on graphical user interface 208 after the user has selected “Add More Music” menu choice 402 .
  • “Add More Music” panel features entry field 502 . Entry field 502 enables the user to selectively modify the input seed by entering, for example, another artist name or song name (in addition to the artist name, song name or other input seed previously entered) and then selecting “Add” button 504 (if the user does not desire to selectively modify the input seed, then the user selects “Close” button 506 ).
  • the additional artist name or song name is then factored into the comparison between the input seed and songs contained in the “MUSIC GENOME PROJECT” database.
  • One way to factor the additional artist name or song name into the comparison is to utilize confidence and weighting factors to assign, for example, more or less weight to the musicological attributes of the additional artist name or song name in view of the initial input seed.
  • “Add More Music” panel 500 disappears and graphical user interface 208 proceeds to present and provide content corresponding to the modified input seed in accordance with FIGS. 2 a and 2 b . In doing so, “Station 1 ” button 308 may appear differently to reflect the modified input seed.
  • “Email This Station” menu choice 404 enables the user to selectively provide a station, and thus a playlist, to another user.
  • FIG. 6 depicts “Email This Station” panel 600 that appears on graphical user interface 208 after the user has selected “Email This Station” menu choice 404 .
  • “Email This Station” panel 600 features station field 602 , email field 604 and message field 606 .
  • station field 602 When selected by the user, enables the user to select a station to selectively provide to another user.
  • the stations available to selectively appear on a drop-down menu and may include stations created by the user, such as the station corresponding to “Station 1 ” button 308 ( FIG. 3 c ), or other stations.
  • Email field 604 enables the user to enter an email address corresponding to another user for which the user desires to selectively provide a station.
  • Message field 606 enables the user to provide a message (such as regular text or HTML) to the user for which the station has been selectively provided.
  • Playlist server 108 prepares an email including the information entered in message field 606 to the recipient user utilizing SMTP or other common protocols.
  • the return address of the email corresponds to the email address provided by the user upon registration with the music discovery service.
  • the email further includes a hypertext link to the URL of the music discovery service.
  • the hypertext link includes a command line argument of an identifier corresponding to the station the user desires to selectively provide. If the recipient is already registered with the music discovery service, the station is automatically provided.
  • the hypertext link will direct the recipient to graphical user interface 208 as if the recipient were the anonymous registrant. If the user does not desire to selectively provide a station, and thus a playlist, to another user, then the user selects “Cancel” button 610 .
  • “Edit This Station” menu choice 406 enables the user to, among other things, selectively modify feedback about a content object such as a song.
  • FIGS. 7 a -c depict “Edit This Station” panel 700 that appears after the user has selected “Edit This Station” menu choice 406 .
  • “Edit This Station” panel 700 features station title 702 , which displays the name of the station (such as the station corresponding to “Station 1 ” button 308 ) that is being edited.
  • “Edit This Station” panel 700 also features “Items You Added” panel 704 , “Songs You Liked” panel 706 and “Songs You Didn't Like” panel 708 . The user may access each of these panels by selecting tab 710 that corresponds to the appropriate panel.
  • “Items You Added” panel 704 features song name text 712 and/or artist name text 714 corresponding to selective modifications of the input seed corresponding to the current station.
  • song name text 712 and artist name text 714 respectively correspond to a song and artist previously entered by the user in order to selectively modify the input seed.
  • the user may remove, for example, a song that had previously selectively modified the input seed by selecting “Remove” button 716 .
  • graphical user interface 208 will no longer present and provide content corresponding to the modified input seed. Instead, graphical user interface 208 will proceed to present and provide content corresponding to, for example, the initial input seed, or to the input seed as selectively modified by entry of artist 714 .
  • “Songs You Liked” panel 706 features, for example, song name text 718 (or artist name text) corresponding to selective feedback that the user has provided about a song.
  • song name text 712 corresponds to a song for which the user has previously selectively provided positive feedback.
  • “Songs You Didn't Like” panel 708 features, for example, song name text 722 (or artist name text) corresponds to a song for which the user has previously selectively provided negative feedback.
  • the user may delete the feedback previously provided by selecting “Remove” button 720 . Thereafter, when the song is provided, graphical user interface 208 will no longer display feedback indicator 336 ( FIG. 3 d ). Multiple songs and/or artists may be listed on “Items You Added” panel 704 , “Songs You Liked” panel 706 or “Songs You Didn't Like” panel 708 . Moreover, the feedback about the song will no longer be utilized in connection with generating playlists.
  • station pop-up menu 400 also features “Rename This Station” menu choice 408 and “Delete This Station” menu choice 410 .
  • “Rename This Station” menu choice 408 enables the user to selectively provide an edited name for, for example, the station that corresponds to “Station 1 ” button 308 .
  • “Delete This Station” menu choice 410 enables the user to remove a station from graphical interface 208 .
  • FIG. 8 depicts, in accordance with an embodiment of the invention, content pop-up menu 800 , which is generated in response to a user selecting a component of a graphical element, such as content art 332 of second graphic element 326 .
  • Content pop-up menu 800 includes menu choices such as “Why Did You Play This Song” menu choice 802 , “Make a New Station from This Song” menu choice 804 , “Buy This Song” menu choice 806 and “Buy This Album” menu choice 808 .
  • Other embodiments of the invention may have fewer, additional or alternative menu choices.
  • FIG. 9 depicts “Information” panel 900 that appears on graphical user interface 208 after the user has selected, for example, “Why Did You Play This Song” menu choice 802 .
  • “Information” panel 900 features information, such as information provided in information text 902 , about the song or other content object currently being provided to the user.
  • “Information” panel 900 may include information relating to a characteristic or focus trait of the song or other content object.
  • “Information” panel 900 may also include information relating to background knowledge about the song, the artist who created the song or other relevant information. To the extent the user no longer desires to review the information, the user selects “Close” button 904 and information panel 900 disappears.
  • “Make a New Station from This Song” menu choice 804 facilitates the presenting of content to a user in accordance with the present invention.
  • FIG. 10 depicts “Create New Station” panel 1000 that appears on graphical user interface 208 after the user has selected, for example, “Make a New Station from This Song” menu choice 804 .
  • “Create New Station” panel 1000 features input seed field 1002 and “Create” button 1004 .
  • input seed field 1002 is automatically filled with the song name corresponding to the song that was provided when content pop-up menu 800 was initially selected.
  • input seed field 1002 is empty and awaits the entry of a song name by the user.
  • a station is automatically created in graphical user interface 208 after the user has selected “Make a New Station from This Song” menu choice 804 .
  • the user selects “Close” button 1006 .
  • “Create New Station” panel 1000 disappears and is replaced on the display by graphical user interface 208 .
  • content pop-up menu 800 features “Buy This Song” menu choice 806 and “Buy This Album” menu choice 808 . If the user selects “Buy This Song” menu choice 806 , then the selective purchase of the song (or other content object) is enabled.
  • One way to enable the selective purchase of the song is to hyperlink “Buy This Song” menu choice 806 to a web site such as the “iTunes” web site from Apple Computer Corp. that offers songs for sale.
  • the hyperlink may include a general URL as well as a parameter specifying the exact song for purchase. If the user selects “Buy This Album” menu choice 808 , then the selective purchase of the album (or other content object) is enabled.
  • One way to enable the selective purchase of the album is to hyperlink “Buy This Album” menu choice 808 to a web site such as the web site of Amazon.com, which sells albums.
  • the hyperlink may include a general URL as well as a parameter specifying the exact song for purchase.
  • Content pop-up menu 800 also includes menu choices such as “I Like It” menu choice 810 and “I Don't Like It” menu choice 812 .
  • “I Like It” menu choice 810 and “I Don't Like It” menu choice 812 enable the user to selectively provide, respectively, positive or negative feedback about the current song or other content object. If the user selects “I Like It” menu choice 810 , then feedback indicator 336 in the shape of, for example, a “thumbs-up” sign is displayed on graphic user interface 208 ( FIG. 3 d ). If the user selects “I Don't Like It” menu choice 812 , then feedback indicator 336 in the shape of, for example, a “thumbs-down” sign is displayed on graphic user interface 208 ( FIG. 3 d ). Other types of feedback, such as “Don't play this song for awhile” may also be selectively provided. As stated previously, feedback may be used to customize and enhance playlists and other aspects of the user experience.
  • Content pop-up menu 800 further includes “Add to Favorites” menu choice 814 .
  • “Add to Favorites” menu choice 814 enables the user to selectively associate the song or other content object with a favorites list.
  • FIG. 11 depicts “Favorites” display 1100 .
  • “Favorites” display 1100 may appear, for example, as a panel in graphical user interface 208 or as a separate web page provided by playlist server 108 . Another way for the user to access “Favorites” display 1100 is by selecting “Favorites” button 210 ( FIG. 2 a ).
  • “Favorites” display 1100 keeps track of songs that the user has identified as good or otherwise significant.
  • “Favorites” display 1100 features management icons 1102 and 1104 , song text 1106 , artist text 1108 and station text 1110 .
  • Management icons 1102 and 1104 enable the user to remove and otherwise manipulate songs listed in the favorites list in “Favorites” display 1100 .
  • Song text 1106 and artist text 1108 provide information about the song that has been selectively associated with the “Favorites” list.
  • Station text 1110 provides the name of the station, such as the station corresponding to the “Station 1 ” button 308 , from which the song was selectively associated with the “Favorites” list.
  • “Favorites” display 1100 also features date 1112 , album purchase icon 1114 and song purchase item 1116 .
  • Date 1112 provides information as to when the song was selectively associated with the “Favorites” list.
  • Album purchase icon 1114 enables the selective purchase of the album (or other content object) from which the song originates.
  • One way to enable the selective purchase of the album is to hyperlink album purchase icon 1114 to a web site such as the web site of Amazon.com, which sells albums.
  • Song purchase icon 1116 enables the selective purchase of the song (or other content object).
  • One way to enable the selective purchase of the song is to hyperlink song purchase icon 1116 to a web site such as the “iTunes” web site from Apple Computer Corp. that offers songs for sale.
  • FIG. 12 depicts a flow diagram overview of a method for presenting and providing content to a user 1200 that can be executed in connection with, for example, the system depicted in FIG. 1 .
  • the user is enabled to selectively provide an input seed.
  • the input seed may be a song name such as “Paint It Black” or even a group of songs such as “Paint It Black” and “Ruby Tuesday.”
  • the input seed may be an artist name such as “Rolling Stones.”
  • Other types of input seeds could include, for example, genre information such as “Classic Rock” or era information such as “ 1960 s.”
  • the input seed is sent to playlist server 108 ( FIG. 1 ) in order to perform the subsequent generation of a playlist. Encryption and other security methods may be used to protect communications between playlist server 108 , content server 118 and/or terminals 102 , 104 and 106 .
  • a playlist is first generated as a result of a comparison between the input seed and a plurality of database items.
  • the input seed is received from terminals 102 , 104 and 106 and the playlist is generated on playlist server 108 .
  • One or more embodiments of the invention utilize the “MUSIC GENOME PROJECT” database, which is a large database of records, each describing a single piece of music and an associated set of search and matching functions that operate on that database.
  • the matching engine effectively calculates the distance between a source song and the other songs in the database and then sorts the results to yield an adjustable number of closest matches.
  • each song is described by a set of characteristics, or “genes”, or more that are collected into logical groups called “chromosomes.”
  • the set of chromosomes make up the genome.
  • One of these major groups in the genome is the “Music Analysis” Chromosome. This particular subset of the entire genome is sometimes referred to as “the genome.”
  • Each gene can be thought of as an orthogonal axis of a multi-dimensional space and each song as a point in that space. Songs that are geometrically close to one another are “good” musical matches. To maximize the effectiveness of the music matching engine, we maximize the effectiveness of this song distance calculation.
  • a given song “S” is represented by a vector containing approximately 150 genes.
  • Each gene corresponds to a characteristic of the music, for example, gender of lead vocalist, level of distortion on the electric guitar, type of background vocals, etc.
  • rock and pop songs have 150 genes
  • rap songs have 350
  • jazz songs have approximately 400.
  • Other genres of music, such as world and classical have 300-500 genes. The system depends on a sufficient number of genes to render useful results.
  • Each gene “s” of this vector has a value of an integer or half-integer between 0 and 5. However, the range of values for characteristics may vary and is not strictly limited to just integers or half-integers between 0 and 5.
  • the distance is better calculated as a sum that is weighted according to each gene's individual significance. Taking this into account, the revised distance can be calculated as follows:
  • the weighting vector is initially established through empirical work done, for example, by a music team that analyzes songs.
  • the weighting vector can be manipulated in various ways that affect the overall behavior of the matching engine. This will be discussed in more detail later in this document.
  • the data represented by many of the individual genes is not linear. In other words, the distance between the values of 1 and 2 is not necessarily the same as the distance between the values of 4 and 5.
  • the introduction of scaling functions f(x) may adjust for this non-linearity. Adding these scaling functions changes the matching function to read:
  • g(x) is simply x 2 , but it could become x 3 for example if it was preferable to prioritize songs with many small differences over ones with a few large ones.
  • Focus matching allows the end user of a system equipped with a matching engine to control the matching behavior of the system.
  • Focus traits may be used to re-weight the song matching system and refine searches for matching songs to include or exclude the selected focus traits.
  • Focus Traits are the distinguishing aspects of a song.
  • Triggering rules are applied to each of the possible focus traits to discover which apply to the song in question. These rules may trigger a focus trait when a given gene rises above a certain threshold, when a given gene is marked as a definer, or when a group of genes fits a specified set of criteria.
  • the identified focus traits (or a subset) are presented on-screen to the user. This tells the user what elements of the selected song are significant.
  • An end user can choose to focus a match around any of the presented traits.
  • the matching engine modifies its weighting vector to more tightly match the selection. This is done by increasing the weights of the genes that are specific to the Focus Trait selected and by changing the values of specific genes that are relevant to the Trait. The resulting songs will closely resemble the source song in the trait(s) selected.
  • the weighting vector can also be manipulated for each end user of the system. By raising the weights of genes that are important to the individual and reducing the weights of those that are not, the matching process can be made to improve with each use.
  • the matching engine is capable of matching songs. That is, given a source song, it can find the set of songs that closely match it by calculating the distances to all known songs and then returning the nearest few. The distance between any two songs is calculated as the weighted Pythagorean sum of the squares of the differences between the corresponding genes of the songs.
  • the source group could represent the collected works of a single artist, the songs on a given CD, the songs that a given end user likes, or analyzed songs that are known to be similar to an unanalyzed song of interest. Depending on the makeup of the group of songs, the match result has a different meaning to the end user but the underlying calculation should be the same.
  • This functionality provides a list of songs that are similar to the repertoire of an artist or CD. Finally, it will allow us to generate recommendations for an end user, purely on taste, without the need for a starting song.
  • FIG. 13 illustrates two songs.
  • the song on the right is a better match to the set of source songs in the center.
  • the virtual “center” is defined to be a song vector whose genes are the arithmetic average of the songs in the original set. Associated with this center vector is a “deviation” vector that represents the distribution of the songs within the set.
  • An individual gene that has a very narrow distribution of values around the average will have a strong affinity for the center value.
  • a gene with a wide distribution on the other hand, will have a weak affinity for the center value.
  • the deviation vector will be used to modify the weighing vector used in song-to-song distance calculations. A small deviation around the center means a higher net weighting value.
  • the center-deviation vector pair can be used in place of the full set of songs for the purpose of calculating distances to other objects.
  • Target song vectors T ( t 2 , . . . t n )
  • Center vector of the source group C ( ⁇ 1 , ⁇ 2 , . . . , ⁇ n )
  • Standard deviation vector of the source group D ( ⁇ 1 , ⁇ 2 , . . . ⁇ n )
  • weighting vectors may be appropriate for multi-song matching of this sort. Different multi-song weighting vector may be established, or the (0.5) 2 constant may be modified to fit with empirically observed matching results.
  • Groups with a coherent, consistent set of tracks will have both a known center vector and a tightly defined deviation vector.
  • This simple vector pair scheme will breakdown, however, when there are several centers of musical style within the collection. In this case we need to describe the set of songs as a set of two or more vector pairs.
  • the song group can be described with two vector pairs. By matching songs to one OR the other of the vector pairs, we will be able to locate songs that fit well with the set. If we were to try to force all of the songs to be described by a single pair, we would return songs in the center of the large ellipse that would not be well matched to either cluster of songs.
  • cluster size there will be a small number of such clusters, each with a large number of closely packed elements. We can then choose to match to a single cluster at a time. In applications where we are permitted several matching results, we can choose to return a few from each cluster according to cluster size.
  • FIG. 16 shows a more detailed flow diagram for one or more embodiments of this step.
  • characteristics that correspond to the input seed are identified.
  • characteristics may include, for example, gender of lead vocalist, level of distortion on the electric guitar, type of background vocals, etc. Characteristics may also include, for example, other types of musicological attributes such as syncopation, which is a shift of accent in a musical piece that occurs when a normally weak beat is stressed. In one or more embodiments of the invention, such characteristics are retrieved from one or more items corresponding to the input seed in a Music Genome Project database.
  • FIG. 17 shows a more detailed flow diagram for one embodiment of the
  • Identify Characteristics step 1604 ( FIG. 16 ). As indicated previously, “Identify Characteristics” step 1604 as well as all of the other steps in FIG. 16 , can be executed on, for example, the servers in FIG. 1 .
  • database 112 in FIG. 1 which may be a Music Genome Project database, is accessed to first identify whether the input seed is an item in database 112 .
  • the user may be asked for more information in an attempt to determine, for example, whether the input seed was inputted wrong (e.g., “Beetles” instead of “Beatles”) or whether the input seed goes by another name in database (e.g., “I feel fine” instead of “She's in love with me”).
  • close matches to the input seed may be retrieved from the database and displayed to the user for selection.
  • the input seed is then categorized.
  • the input seed is categorized as either a “Song Name” or “Artist Name.”
  • Such categorization is realized by, for example, retrieving “Song Name” or “Artist Name” information associated with the input seed from the database.
  • such categorization is realized by asking the user whether the input seed is a “Song Name” or “Artist Name.”
  • “Retrieve Characteristics” step 1704 is executed.
  • “Retrieve Characteristics” step 1704 a song vector “S” that corresponds to the song is retrieved from the database for later comparison to another song vector.
  • the song vector contains approximately 150 characteristics, and may have 400 or more characteristics:
  • Each characteristic “s” of this vector has a value selected from a range of values established for that particular characteristic.
  • the value of the “syncopation” characteristic may be any integer or half-integer between 0 and 5.
  • the value of the syncopation characteristic for most “Pink Floyd” songs is 2 or 2.5.
  • the range of values for characteristics may vary and is not limited to just integers or half-integers between 0 and 5.
  • Average step 1706 is executed.
  • song vectors S 1 to S n which each correspond to one of n songs in the database by the artist that is the subject of the input seed, are retrieved.
  • song vectors S 1 to S n could correspond to one of n songs in the database on a particular album by the artist.
  • Center vector C ( ⁇ 1 , ⁇ 2 , . . . ⁇ n )
  • ⁇ 1 ( s 1,1 +s 2,1 + . . . s n,1 )/ n
  • center vector “C” is then used for later comparison to another song vector as a representation of, for example, the average of all songs by the artist.
  • center vector “C 1 ” corresponding to a first artist may be compared to center vector “C 2 ” corresponding to a second artist.
  • “assign confidence factor” step 1708 is executed.
  • a deviation vector “D” is calculated:
  • Deviation Vector D ( ⁇ 1 , ⁇ 2 , . . . ⁇ n )
  • the value of the syncopation characteristic for most “Pink Floyd” songs is 2 or 2.5, which results in a smaller standard deviation value (e.g., 0.035) than if a standard deviation value were calculated for a characteristic having more divergent values (e.g., if the value of the syncopation characteristic for all songs by Pink Floyd was more widely dispersed between 0 and 5).
  • the values of the deviation vector serve as “confidence factors” that emphasize values in the virtual song vector depending on their respective reliabilities.
  • One way to implement the confidence factor is by multiplying the result of a comparison between the center vector and another song vector by the inverse of the standard deviation value.
  • the confidence factor could have a value of 0.25/( ⁇ i ⁇ 2+0.25). The “0.25” is put into the equation to avoid a mathematically undefined result in the event ⁇ i ⁇ 2 is 0 (i.e., the confidence factor avoids “divide by zero” situations).
  • Focus Traits identifies focus traits based on the values of characteristics of song vector (or virtual song vector) S.
  • focus traits are specific combinations of characteristics (or even a single notable characteristic) representing significantly discernable attributes of a song.
  • focus traits are the kernel of what makes one song actually sound different, or like, another song.
  • Focus traits may be created and defined in many ways, including by having trained musicologists determine what actually makes one song sound different from another, or else having users identify personal preferences (e.g., receiving input from a user stating that he/she likes songs with male lead vocals).
  • Exemplary focus traits include “male lead vocal” or “Middle Eastern influence.” There can be 1, 10, 1000 or more than 1000 focus traits, depending on the desired complexity of the system.
  • a set of rules known as “triggers” is applied to certain characteristics of song vector S to identify focus traits.
  • the trigger for the focus trait “male lead vocal” may require the characteristic “lead vocal present in song” to have a value of 5 on a scale of 0 to 5, and the characteristic “gender” to also have a value of 5 on a scale of 0 to 5 (where “0” is female and “5” is male). If both characteristic values are 5, then the “male lead vocal” focus trait is identified. This process is repeated for each focus trait. Thereafter, any identified focus traits may be presented to the user through the user interface.
  • Weighting Factor Assignment step 1608 is executed.
  • weighting factor assignment step 1608 , comparative emphasis is placed on some or all of focus traits by assigning “weighting factors” to characteristics that triggered the focus traits. Alternatively, “weighting factors” could be applied directly to certain characteristics.
  • weighting factors are assigned based on a focus trait weighting vector W, where w 1 , w 2 and w n correspond to characteristics s 1 , s 2 and s n of song vector S.
  • Weighting Vector W ( w 1 , w 2 , w 3 , . . . , w n )
  • weighting vector W can be implemented into the comparison of songs having and song vectors “S” and “T” by the following formula:
  • one way to calculate weighting factors is through scaling functions. For example, assume as before that the trigger for the focus trait “male lead vocal” requires the characteristic “lead vocal present in song” to have a value of 5 on a scale of 0 to 5, and the characteristic “gender” to also have a value of 5 on a scale of 0 to 5 (where “0” is female and “5” is male).
  • weighting vector W is populated with weighting factors of, for example, 100 for w 1 and w 2 .
  • weighting vector W could receive different weighting factors for w 1 and w 2 (e.g., 10 and 1000, respectively).
  • weighting vector W may multiply every difference in characteristics that trigger a particular focus trait by 100. Accordingly the equation becomes w 1 (s 1 ⁇ t 1 ) ⁇ 2+w 2 (s 2 ⁇ t 2 ) ⁇ 2, which is 100(5 ⁇ 5) ⁇ 2+100(5 ⁇ 0) ⁇ 2, or 2500.
  • the distance of 2500 is much further away than 25 and skews the result such that songs having a different gender of the lead vocalist are much less likely to match.
  • a weighting vector value of 1,000,000 in this circumstance would effectively eviscerate any other unweighted matches of characteristics and means that, in most circumstances, two songs would never turn up as being similar.
  • one or more values of the weighting vector may be assigned based on preferences of the user.
  • a user could identify a “male lead vocal” as being the single-most important aspect of songs that he/she prefers.
  • a weighting vector value of 10,000 may be applied to the comparison of the characteristics associated with the “male lead vocal” focus trait. As before, doing so in one embodiment of the invention will drown out other comparisons.
  • one weighting vector is calculated for each focus trait identified in a song. For example, if 10 focus traits are identified in a song (e.g., “male lead vocalist” and 9 other focus traits), then 10 weighting vectors are calculated. Each of the 10 weighting vectors is stored for potential use during “Compare Weighted Characteristics” step 1610 . In one embodiment of the invention, users can select which focus traits are important to them and only weighting vectors corresponding to those focus traits will be used during “Compare Weighted Characteristics” step 1610 . Alternatively, weighting vectors themselves could be weighted to more precisely match songs and generate playlists.
  • the closest songs are selected for the playlist based on the comparison performed in “Compare Weighted Characteristics” step 1610 .
  • the 20 “closest” songs are preliminary selected for the playlist and placed into a playlist set. Individual songs are then chosen for the playlist.
  • One way to choose songs for the playlist is by random selection. For example, 3 of the 20 songs can be randomly chosen from the set.
  • another song by the same artist as the input seed is selected for the playlist before any other songs are chosen from the playlist.
  • One way to do so is to limit the universe of songs in the database to only songs by a particular artist and then to execute the playlist generating method.
  • a plurality of sets of closest songs are obtained. For example, if a song has 10 focus traits and the 20 closest songs are preliminarily selected for the playlist, then 10 different sets of 20 songs each (200 songs total) will be preliminarily selected. Songs can be selected for the playlist from each of the sets by, for example, random selection. Alternatively, each set can have songs be selected for the playlist in order corresponding to the significance of a particular focus trait.
  • rules may be implemented to govern the selection behavior. For example, aesthetic criteria may be established to prevent the same artist's songs from being played back-to-back after the first two songs, or to prevent song repetition within 4 hours.
  • regulatory criteria may be established to comply with, for example, copyright license agreements (e.g., to prevent the same artist's songs from being played more than 4 times in 3 hours).
  • copyright license agreements e.g., to prevent the same artist's songs from being played more than 4 times in 3 hours.
  • a history of songs that have been played may be stored along with the time such songs were played.
  • songs are selected for the playlist from one or more playlist sets according to random selection, aesthetic criteria and/or regulatory criteria.
  • focus traits can be ranked (e.g., start with all selected songs from the playlist set deriving from the “male lead vocal” focus trait and then move to the next focus trait).
  • the user can emphasize or de-emphasize particular playlist sets. If, for example, a user decides that he/she does not like songs having the focus trait of “male lead vocal,” songs in that playlist set can be limited in the playlist.
  • a number of songs are selected from the Set List and played in sequence as a Set. Selection is random, but limited to satisfy aesthetic and business interests, (e.g. play duration of a particular range of minutes, limits on the number of repetitions of a particular Song or performing artist within a time interval).
  • a typical Set of music might consist of 3 to 5 Songs, playing for 10 to 20 minutes, with sets further limited such that there are no song repetitions within 4 hours and no more than 4 artist repetitions within 3 hours.
  • the playlist features identifiers that correspond to, for example, song names.
  • the identifiers may be index fields or other handles for content database 120 on content server 118 .
  • playlist server 108 may send an identifier corresponding to the input seed to the user at terminal 102 , 104 or 106 .
  • playlist server 108 may, for example, send an identifier corresponding to a song that is the closest match to the “center vector.”
  • a set of identifiers may be sent to terminal 102 , 104 or 106 (or to multiple terminals) at once.
  • the player on terminal 102 , 104 or 106 proceeds to associate a graphic element (such as first graphic element 314 in FIG. 3 ) with the identifier.
  • content server 118 may store song name 316 , artist name 318 and content art 320 in connection with a corresponding content object in content database 120 .
  • the player on terminal 102 , 104 or 106 may request song name 316 , artist name 318 and content art 320 that corresponds to the input seed or identifier from content server 118 .
  • Content server 118 then provides, in encrypted form, song name 316 , artist name 318 and content art 320 to the player on terminal 102 , 104 or 106 .
  • first graphic element 314 appears in graphical user interface 208 as discussed previously.
  • Song name 316 , artist name 318 and content art 320 may be provided within first graphic element 314 .
  • a content object corresponding to the identifier or input seed is provided.
  • the player in terminal 102 , 104 or 106 may send the identifier received from playlist server 108 to content server 118 .
  • content server 118 may provide a content object corresponding to the identifier to the player on terminal 102 , 104 or 106 and thus to the user.
  • Content server 118 may provide a content object to the user in several ways.
  • content server 118 may stream content object to the user through well-known streaming techniques and protocols such as User Datagram Protocol (UDP), Real Time Transport Protocol (RTP), Real Time Streaming Protocol (RTSP), Real Time Control Protocol (RTCP) and Transmission Control Protocol (TCP).
  • UDP User Datagram Protocol
  • RTP Real Time Transport Protocol
  • RTSP Real Time Streaming Protocol
  • RTCP Real Time Control Protocol
  • TCP Transmission Control Protocol
  • content server 118 may provide a content object to the user through downloading.
  • the content object is downloaded fully to terminal 102 , 104 or 106 before it is provided to the user.
  • the content object may be provided to the user through a hybrid method of streaming and downloading.
  • content server 118 may provide a content object at a rate of 10 to 20 times that of the playback rate. Portions of the content object that have not been played are cached in memory on terminal 102 , 104 or 106 .
  • a content object corresponding to the identifier or input seed is provided.
  • the player in terminal 102 , 104 or 106 may send the identifier received from playlist server 108 to content server 118 .
  • content server 118 may provide a content object corresponding to the identifier to the player on terminal 102 , 104 or 106 and thus to the user.
  • “Obtain Feedback” step 1212 the user selectively provides feedback about a content object through graphical user interface 208 in the manner discussed previously.
  • feedback that has been selectively provided by the user is sent to playlist server 108 .
  • the playlist may be modified as discussed previously.
  • the user may selectively provide feedback that is negative about a song with a focus trait of “male lead vocal.”
  • a new playlist is generated by playlist server 108 (i.e., the existing playlist is modified) that accounts for the negative feedback.
  • a weighting value or scaling function corresponding to the focus trait of “male lead vocal” may be adjusted such that songs having strong focus traits of “male lead vocal” are less likely to match with the input seed originally provided by the user.
  • the user may selectively provide feedback that he or she does not like “jazz” music.
  • “Jazz” may be a characteristic stored with regard to various songs in database 112 . A weighting value of 1/1,000,000,000 is then assigned to the characteristic “jazz,” which means that a match between the input seed and “jazz” songs is unlikely to result from a comparison of the input seed and database items. Accordingly, the playlist will be modified to remove jazz songs.
  • FIG. 18 shows a flow diagram of a method of generating a playlist of content objects in accordance with aspects of the invention.
  • Obtain Feedback step 1212 in FIG. 12
  • feedback about a content object may be obtained from a user and sent to playlist server 108 .
  • Store Feedback step 1802
  • feedback from the user may be stored in memory on playlist server 108 (e.g., in database 112 )( FIG. 1 ).
  • the feedback may be stored in memory on a content server 118 .
  • the feedback may be stored on any computing device containing memory that is accessible from network 110 .
  • a separate feedback server and/or database may be beneficial in some configurations familiar to those of skill in the art.
  • the feedback stored in “Store Feedback” step 1802 may be positive, negative, or neutral. Numerous examples of negative feedback and positive feedback have been provided above. Meanwhile, in accordance with aspects of the invention, instances when the user refuses to give feedback or simply fails to give feedback are stored as neutral (or null) feedback. Neutral feedback tracks the number of occurrence of the absence of feedback. Therefore, in one aspect of the invention, the total number of times a particular song is played (i.e., “spins” of that song) is equal to the sum of the quantity of neutral feedbacks, positive feedbacks, and negative feedbacks stored for that song.
  • contextual information relating to the feedback is stored in “Store Contextual Information” step 1804 . Similar to the storing of feedback, the contextual information may also be stored in memory on playlist server 108 . Alternatively, the contextual information may be stored in memory on a content server 118 . One of skill in the art will appreciate that the contextual information may be stored on any computing device containing memory that is accessible from network 110 . For example, it may be beneficial in some configuration to setup a separate server and/or database for feedback and contextual information.
  • FIG. 19 which is discussed in detail below, depicts an example of feedback and contextual information stored together in a database.
  • contextual information include, but are not limited to, the station on which the current song is playing, the song played immediately prior to the current song, the artist of the song playing immediately prior to the current song, one or more musicological characteristics of the prior song(s), the artist of the current song, the album containing the current song, time of day, day of week, calendar day (i.e., date), location of user (e.g., work, home, car, zip code, country, etc.), and user demographics (e.g., age, gender).
  • Some contextual information may be related to the current song in the context of an extrinsic factors, such as location of user (i.e., geography), user demographics, and/or temporal (e.g., time of day, day of week, calendar date). Meanwhile, some contextual information may be related to the current song in the context of at least one of the following: particular artist, particular album, and/or particular station. In addition, some contextual information may be related to the current song in the context of a second content object (e.g., an immediately prior song). Moreover, some contextual information may be related to the current song in the context of characteristics of a second content object (e.g., characteristics of an immediately prior song such as treble, bass, vocals, etc.). Examples of each of the various types of contextual information will be clear to one of skill in the art after thorough review of the entirety of the disclosure.
  • contextual information relating to that feedback is automatically obtained and stored in memory.
  • the particular station e.g., “Pet Shop Boys” Station
  • the particular artist e.g., “Pet Shop Boys”
  • the time of day e.g., 9 a.m.
  • day of week e.g., Monday
  • the fact that the user giving feedback was a 28-year old male located in Virginia may be stored as contextual information relating to the feedback.
  • stored feedback and contextual information is retrieved in “Retrieve Feedback and Contextual Information” step 1806 .
  • the stored feedback and contextual information related to the “Pet Shop Boys” Station is retrieved and used in generating a playlist of content objects (e.g., songs).
  • the feedback and contextual information are used in generating a playlist only after the quantity of feedback has exceeded a statistical threshold. In examples where the threshold has not been met, the generation of playlists may be similar to step 1204 .
  • the feedback and contextual information are used in playlist generation and/or optimization only after at least one hundred spins of a song in a particular context (e.g., the song “With or Without You” playing on “Pet Shop Boys” Station).
  • the statistical threshold value may be set to 1,000 spins or 10,000 spins.
  • a larger threshold value may provide a higher level of confidence in the statistical reliability of the information collected, but may delay the resulting playlist optimizations.
  • a playlist is generated by playlist server 108 that accounts for negative, positive, and neutral feedback provided in that particular context. Unlike the generation and modification of a playlist in step 1204 , this step uses feedback with contextual information to deliver an enhanced playlist.
  • the context in which a song is played may be a station having an input seed of artist “U2” (i.e., in the context of songs played on the “U2” Station).
  • the feedback and contextual information stored for that particular context is depicted by the appropriate portion of database table 1900 depicted in FIG. 19 .
  • Database table 1900 in FIG. 19 stores contextual information and feedback provided by users in the context of stations.
  • database table 1900 shows only some of the entries stored for a “U2” Station (i.e., a station based on an artist input seed) and a “Sweet Child of Mine” Station (i.e., a station based on a song input seed).
  • Database table 1900 contains ten entries for each station, where each entry corresponds to a song that was played in the context of the “Sweet Child of Mine” Station or “U2” Station.
  • the number of entries in database table 1900 may dynamically grow as different songs are played on that station and added to database table 1900 .
  • database table 1900 may be prepopulated with an entry corresponding to every available song.
  • prepopulating database table 1900 may undesirably result in extremely large database tables in memory.
  • database entry 1902 corresponds to the feedback provided by users listening to the “U2” Station in response to hearing artist U2's song “With or Without You.”
  • the “Artist Name” column contains “U2”
  • the “Song Name” column contains “With or Without You”.
  • Other relevant portions of database table 1900 for entry 1902 are as follows: the “Spins” column contains 31,271 (i.e., the total number of times the song “With or Without You” has played on a “U2” Station), the “Positive” column contains the number of times a user gave positive feedback (e.g., by selecting “I Like It” menu choice 810 ( FIG.
  • the “Negative” column contains the number of times a user gave positive feedback (e.g., by selecting “I Don't Like It” menu choice 812 ( FIG. 8 ) for the song) for the song. Meanwhile, as explained earlier, the quantity of neutral feedback can be obtained by simple arithmetic (i.e., Spins minus Positive minus Negative).
  • the “Feedback” column is a total of positive and negative feedback.
  • the “Feedback” column may be calculated using other columns in database table 1900 and need not be present in all embodiments.
  • database entry 1902 contains a “Song Q” column that stores values indicating the suitability of its respective content object in that particular context.
  • the suitability quotient “Q” may be expressed as a numerical value between negative one and positive one.
  • Negative “Q” values may be used to indicate that a particular content object (e.g., the song “With or Without You”) is less suitable in the particular context (e.g., on “Sweet Child of Mine” Station that song receives a negative 0.064 “Q” value).
  • a positive “Q” value may be used to indicate that a particular content object (e.g., the song “With or Without You”) is more suitable in the particular context (e.g., on “U2” Station that song receives a positive 0.045 “Q” value).
  • the “Q” value may be recalculated and updated in database table 1900 each time a song is played in the context of a particular station.
  • feedback and contextual information may be collected and the corresponding “Q” values recalculated and updated only at regular intervals (e.g., 30 minutes or after 100 new spins).
  • the suitability quotient “Q” for a content object in a particular context is calculated using the function:
  • F pos is the sum of all positive feedback for the content object
  • F neg is the sum of all negative feedback for the content object
  • total # of spins is the total number of times that a content object was performed in that particular context.
  • database entry 1902 shows that the song “With or Without You” by U2 was played on the “U2” Station 31,271 times. Of those times, the song received positive feedback 1,500 times and negative feedback 96 times.
  • the “Q” value for the song “With or Without You” in the context of the “U2” Station is 0.045 (i.e., 1,500 minus 96, divided by 31,271). Therefore, a positive “Q” value for that song in that context will increase the likelihood that it will be found in the playlist generated in step 1808 .
  • a weighting value or scaling function as discussed previously and that is appropriate for the suitability quotient “Q” value may be used to optimize the comparison performed by the matching algorithm.
  • the matching algorithm was discussed extensively above. For example, according to the illustrative data in database table 1900 under the “U2” Station context, the song “Age of Consent” by New Order has a more favorable suitability quotient “Q” value than the song “We Are One” and is thus more likely to be selected for the playlist during a comparison.
  • a number of content objects may be selected and played in a sequence determined, at least in part, using each song's respective “Q” value in the particular context. For example, according to the illustrative data in database table 1900 under the “U2” Station context, the song “Age of Consent” by New Order has a more favorable suitability quotient “Q” value than the song “We Are One” and is thus more likely to be selected for the playlist during a comparison.
  • the “Q” may be used to optimize the generation of a pleasing sequence of music (e.g., content objects) in a particular context.
  • a number of content objects e.g., songs
  • FIG. 20 depicts another flow diagram of a method of generating a playlist of content objects in accordance with aspects of the invention.
  • an input seed is compared to a first database item corresponding to a first content object (i.e., song).
  • the comparison may use the matching algorithm discussed above to identify a playlist of one or more songs that sound alike (e.g., are closely related to a source song or group of songs based on their characteristics and weighted comparisons of these characteristics). For example, a user may enter the artist “Massive Attack” as an input seed and be presented with one or more database items corresponding to content objects (e.g., the song “With or Without You”).
  • a database item may be simply a unique identifier for a particular song.
  • step 2004 contextual information relating to a first content object (e.g., the song “With or Without You”) is retrieved from memory.
  • a first content object e.g., the song “With or Without You”
  • the suitability quotient “Q” calculated for the song “With or Without You” on the “Massive Attack” Station may be retrieved.
  • the retrieving performed in step 2004 is similar to the retrieving performed in step 1806 .
  • the suitability quotient “Q” may optimize the actual order and/or placement of the songs in the playlist.
  • step 2006 the contextual information relating to the first content object (e.g., the song “With or Without You”) is used to determine whether to select that object for the playlist.
  • Step 2006 which is an enhancement of step 1612 ( FIG. 16 ) uses retrieved contextual information in addition to the comparison algorithm used in step 1612 to select content objects for the playlist.
  • step 2006 may show that the song “With or Without You” corresponds to a negative “Q” value when played after the song “Angel”.
  • Such information may be stored in a table in memory like table 1900 in FIG. 19 .
  • the context in which the feedback was obtained would relate to the prior content object (e.g., the song “Angel”) and not the station, as was the case in a prior example.

Abstract

Systems and methods of generating and modifying a playlist using contextual information are disclosed. For example, a user may provide an input seed such as a song name or artist name. The input seed is compared to database items and a playlist is generated as a result. In some examples, the contextual information is used to enhance the comparison and to select better content objects for a user's playlist. Meanwhile, in some examples the content objects in the playlist may be arranged in an enhanced order using ranking and/or contextual information.

Description

  • This application is a continuation of U.S. patent application Ser. No. 13/108,184, filed May 16, 2011, which is a continuation of U.S. patent application Ser. No. 11/380,563, filed Apr. 27, 2006, now U.S. Pat. No. 7,962,482, which is a continuation-in-part of U.S. patent application Ser. No. 11/279,567, filed Apr. 13, 2006, which is a continuation-in-part of U.S. patent application Ser. No. 11/295,339, filed Dec. 6, 2005, which is a continuation-in-part of U.S. patent application Ser. No. 10/150,876, filed May 16, 2002, now U.S. Pat. No. 7,003,515, which claims priority to provisional U.S. Patent Application Ser. No. 60/291,821, filed May 16, 2001. The entire disclosures of U.S. patent application Ser. Nos. 13/108,184, 11/380,563, 11/279,567, 11/295,339, 10/150,876 and 60/291,821 are hereby incorporated fully by reference.
  • FIELD OF THE EMBODIMENTS OF THE INVENTION
  • Embodiments of the invention are directed to methods and systems for utilizing contextual feedback to generate and modify playlists.
  • BACKGROUND OF THE EMBODIMENTS OF THE INVENTION
  • Automated techniques exist for identifying items that a consumer might enjoy in view of other items the consumer has previously indicated he or she enjoys. Some such techniques compare attributes of items the consumer previously indicated he or she enjoys with attributes of other items to identify items that the consumer might enjoy. Thus, for example, if the consumer enjoys “Dubliners” by James Joyce, “Ulysses” by James Joyce might be identified as another item the consumer may enjoy because both “Dubliners” and “Ulysses” have a common attribute (the author, James Joyce).
  • Other automated techniques utilize collaborative methods to identify items that the consumer might enjoy. For example, consumers who enjoyed “The Da Vinci Code” by Mark Brown might indicate that they also enjoyed “The Catcher in the Rye” by J. D. Salinger. Accordingly, if the consumer indicates that he or she enjoys “The Da Vinci Code,” “The Catcher in the Rye” would be identified as another item the consumer would enjoy because other consumers who enjoyed “The Da Vinci Code” indicated that they enjoyed “The Catcher in the Rye.”
  • One problem with these techniques is that they neglect the context of the attributes or information used to identify items that the consumer might enjoy. For example, consumers who enjoyed “The Da Vinci Code” and who indicated that they also enjoyed “The Catcher in the Rye” might only have enjoyed “The Catcher in the Rye” because they read “Dubliners,” which is written in a similar stream-of-consciousness style, immediately before “The Catcher in the Rye.” If so, then a recommendation such as “If you like ‘The Da Vinci Code,’ then you'll also like ‘The Catcher in the Rye’” may not be helpful to the consumer.
  • This problem also arises where the items are musical compositions. For example, radio station listeners may enjoy the song “House of the Rising Sun” by The Animals after the song “Stairway to Heaven” by Led Zeppelin, but not after the song “Aqualung” by Jethro Tull. However, existing automated techniques are only able to determine, for example, that radio station listeners who like “Stairway to Heaven” and “Aqualung” also like “House of the Rising Sun.” Accordingly, existing automated techniques cannot, for example, help create radio station playlists that account for listener preferences as to the context of a particular musical composition. Accordingly, there exists a need for methods and systems for using contextual information to generate and modify playlists that do account for such listener preferences.
  • BRIEF SUMMARY OF EMBODIMENTS OF THE INVENTION
  • Embodiments of the invention are directed to methods and systems for using contextual information to generate and modify playlists. For example, one embodiment of the invention includes the steps of storing feedback from at least one user about a first content object; storing contextual information relating to the feedback; retrieving the feedback and contextual information; and generating a playlist of content objects using the feedback and contextual information.
  • Another embodiment of the invention includes the steps of comparing an input seed to a first database item, wherein the first database item corresponds to a first content object; retrieving contextual information relating to the first content object; and selecting the first content object for the playlist based on the comparison and the contextual information. In other embodiments of the invention, computer-executable instructions for implementing the disclosed methods are stored as control logic or computer-readable instructions on computer-readable media, such as an optical or magnetic disk.
  • Other details features and advantages of embodiments of the invention will become apparent with reference to the following detailed description and the figures.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 depicts an exemplary operating environment for an embodiment of the invention;
  • FIGS. 2 a and 2 b depict terminal-based displays for presenting and providing content to a user in accordance with embodiments of the invention;
  • FIGS. 3 a-3 d depict in more detail the graphical user interface of FIGS. 2 a and 2 b in various stages of operation and in accordance with an embodiment of the invention;
  • FIG. 4 depicts, in accordance with an embodiment of the invention, a station pop-up menu generated in response to a user selecting a button such as “Station 1button 308 in FIG. 3 c.
  • FIG. 5 depicts, in accordance with an embodiment of the invention, the graphical user interface of FIGS. 2 a and 2 b after a user has clicked the “Add More Music” menu choice 402 of station pop-up menu 400 in FIG. 4;
  • FIG. 6 depicts, in accordance with an embodiment of the invention, the graphical user interface of FIGS. 2 a and 2 b after a user has clicked the “Email This Station” menu choice 404 of station pop-up menu 400 in FIG. 4;
  • FIGS. 7 a-c depict, in accordance with an embodiment of the invention, the graphical user interface of FIGS. 2 a and 2 b in various stages of operation after a user has clicked the “Edit This Station” menu choice 406 of station pop-up menu 400 in FIG. 4;
  • FIG. 8 depicts, in accordance with an embodiment of the invention, a content pop-up menu generated in response to a user selecting a component of a graphical element, such as content art 332 of second graphic element 326.
  • FIG. 9 depicts, in accordance with an embodiment of the invention, an “Information” panel 900 that appears on graphical user interface 208 after the user has selected, for example, “Why Did You Play This Song” menu choice 802.
  • FIG. 10 depicts, in accordance with an embodiment of the invention, a “Create New Station” panel 1000 that appears on graphical user interface 208 after the user has selected, for example, “Make a New Station from This Song” menu choice 804.
  • FIG. 11 depicts a “Favorites” display 1100 in accordance with an embodiment of the invention.
  • FIG. 12 depicts a flow diagram overview of methods for presenting and providing content to a user.
  • FIG. 13 depicts a relationship between different song candidates.
  • FIG. 14 is a graph showing a deviation vector.
  • FIG. 15 graphically depicts a bimodal song group.
  • FIG. 16 shows a flow diagram for one or more embodiments of the “Generate or Modify Playlist” step 1204 in FIG. 12.
  • FIG. 17 shows a flow diagram for one or more embodiments of the “Identify Characteristics” step 1604 in FIG. 16.
  • FIG. 18 shows a flow diagram of a method of generating a playlist of content objects in accordance with aspects of the invention.
  • FIG. 19 depicts feedback and contextual information stored in accordance with aspects of the invention.
  • FIG. 20 shows another flow diagram of a method of generating a playlist of content objects in accordance with aspects of the invention.
  • DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION
  • FIG. 1 depicts a diagram of exemplary system 100 that may be used to implement embodiments of the invention. A plurality of terminals, such as terminals 102, 104 and 106, coupled to playlist server 108 and content server 118 via network 110. In another embodiment, playlist server 108 and content server 118 may be the same server performing all functions of playlist server 108 and content server 118. Terminals 102, 104 and 106, playlist server 108 and content server 118, may include a processor, memory and other conventional electronic components and may be programmed with processor-executable instructions to facilitate communication via network 110 and perform aspects of the invention.
  • One skilled in the art will appreciate that network 110 is not limited to a particular type of network. For example, network 110 may feature one or more wide area networks (WANs), such as the Internet. Network 110 may also feature one or more local area networks (LANs) having one or more of the well-known LAN topologies and the use of a variety of different protocols on these topologies, such as Ethernet, TCP/IP, Frame Relay, Ethernet, FTP, HTTP and the like, is presumed. Moreover, network 110 may feature a Public Switched Telephone Network (PSTN) featuring land-line and cellular telephone terminals, or else a network featuring a combination of any or all of the above. Terminals 102, 104 and 106 may be coupled to network 110 via, for example, twisted pair wires, coaxial cable, fiber optics, electromagnetic waves or other media.
  • In one embodiment of the invention, playlist server 108 contains a database of items 112. Alternatively, playlist server 108 may be coupled to database of items 112. For example, playlist server 108 may be coupled to a “MUSIC GENOME PROJECT” database as described in U.S. Pat. No. 7,003,515. Playlist server 108 may also contain or be coupled to matching engine 114. Matching engine 114 utilizes an associated set of search and matching functions 116 to operate on the database of items 112. In an embodiment of the invention used with the “MUSIC GENOME PROJECT” database, for example, matching engine 114 utilizes search and matching functions implemented in software or hardware to effectively calculate the distance between a source song and other songs in the database (as described here and in U.S. Pat. No. 7,003,515), and then sorts the results to yield an adjustable number of closest matches.
  • In one embodiment of the invention, content server 118 contains a database of content objects 120. Alternatively, content server 118 may be wholly or partially integrated with playlist server 108, or separately coupled to a database of content objects 120. Content server 118 may also contain or be coupled to content engine 122. Content engine 122 utilizes an associated set of management functions 124, such as standard finding, packaging and sending functions, to operate on the database of content objects 122. In one embodiment of the invention, for example, content engine 122 utilizes management functions implemented in software or hardware to control the transmission of content objects by, for example, streaming and/or downloading to terminals 102, 104 and 106.
  • Terminals 102, 104 and 106 feature user interfaces that enable users to interact with server 108. The user interfaces may allow users to utilize a variety of functions, such as displaying information from playlist server 108, requesting additional information from playlist server 108, customizing local and/or remote aspects of the system and controlling local and/or remote aspects of the system. Terminals 102, 104 and 106 can be operated in a client-server configuration to permit a user to retrieve web pages from playlist server 108. Furthermore, any of various conventional web browsers can be used to display and manipulate data on the web pages.
  • FIG. 2 a depicts terminal-based display 200 for presenting and providing content to a user in accordance with an embodiment of the invention. Terminal-based display 200 may comprise, for example, a web browser window 204 displayed on terminal 102 (FIG. 1) running an operating system such as “WINDOWS” from Microsoft Corp. In this embodiment, terminal 102 is configured as the client in a client/server relationship with playlist server 108 and content server 118.
  • A user of terminal 102 establishes a client/server relationship with playlist server 108 by inputting the appropriate URL in address field 206 (in this case, the URL is “http://www.pandora.com”). In response, web page 204 is retrieved from playlist server 108. In this embodiment, web page 204 features graphical user interface 208 (shown in more detail in, e.g., FIG. 3 d), “favorites” button 210, “minimize” button 212, tip 214 and advertisement 216.
  • In this embodiment, the user's selecting of “minimize” button 212 (such as by clicking a mouse button while the mouse pointer is over “minimize” button 212) removes graphical user interface 208 from web page 204 and results in the creation of terminal-based display 220 shown in FIG. 2 b. Terminal-based display 220 presents and provides content to a user in accordance with another embodiment of the invention. Specifically, terminal-based display 220 may comprise, for example, a web browser window 222 featuring graphical user interface 208 without, for example, “favorites” button 210, “minimize” button 212, tip 214 and advertisement 216. Terminal-based display 220 is smaller than terminal-based display 200 and thus better preserves desktop display resources. In a web page replacing web page 204, the user is given the option to return graphical user interface 208 to terminal-based display 200. The user of terminal 102 may discontinue the client/server relationship with playlist server 108 by selecting “close window” button 218. To the extent the user later opens a new web browser window and reestablishes a client/server relationship with playlist server 108, playlist server 108 recognizes the user as a result of well-known schemes such as “cookies” and thus retains any customized user preferences or settings when web page 204 is retrieved and graphical user interface 208 is restarted.
  • In this embodiment, tip 214 enhances the user-friendliness of graphical user interface 208 by providing information to the user regarding how to use graphical user interface 208. For example, tip 214 may state “Use thumbs up/thumbs down to tune your stations. Click here to learn more.” To the extent the user clicks the hypertext link “Click here,” another web page is retrieved providing more detailed information about how to tune stations. Tip 214 may also advertise career opportunities or display other information. In another embodiment, tip 214 may be provided in connection with terminal-based display 220.
  • In this embodiment, advertisement 216 may comprise a standard paid “banner” advertisement for a third party in any configuration on web page 204. Advertisement 216 may generate royalty revenue or other income for the operator. In one embodiment, the type of advertisement 216 presented to the user on web page 204 depends on various criteria, including but not limited to input, feedback and other information provided by the user, the location of the user's IP address, and other information such as the time of day or year.
  • FIGS. 3 a-3 d depict in more detail graphical user interface 208 (FIGS. 2 a and 2 b) in various stages of operation and in accordance with an embodiment of the invention. Graphical user interface 208 is provided through playlist server 108 (FIG. 1) and may be implemented through, for example, Java, JavaScript, XML, Flash or HTML.
  • Turning to FIG. 3 c, graphical user interface 208 features station panel 302 and playlist panel 304. Other embodiments may have more or less panels. Station panel 302 features “Create Station” button 306 and “Station 1,” “Station 2” and “Station 3 buttons 308, 310 and 312.
  • As will be described further below, “Create Station” button 306 initiates the generation of a station (e.g., a station corresponding to “Station 1” button 308) corresponding to an input seed, such as a song name or artist name, selectively provided by the user. The station facilitates the providing of content to the user that, for example, corresponds to a playlist generated as a result of a comparison of the input seed to musicological attributes of other songs. Thus, for example, the user could input “Miles Davis” and a “Miles Davis station” would be created that facilitates the providing of content to the user that corresponds to “Miles Davis” songs or songs that are musicologically similar to songs by “Miles Davis.”
  • In this embodiment, playlist panel 304 visually represents to the user a playlist of content objects such as songs, the first song of which corresponds to first graphic element 314 and the second song of which corresponds to second graphic element 326. First graphic element features corresponding song text 316, artist text 318 and content art 320, while second graphic element 326 features corresponding song text 328, artist text 330 and content art 332. Corresponding song text 316 and 328, as well as corresponding artist text 318 and 330 may additionally comprise hypertext links that provide additional information, such as background knowledge about an artist or song. Corresponding content art 320 and 332 may comprise, for example, a picture of an album cover.
  • Other embodiments of first graphic element 314 or second graphic element 326 may feature additional or fewer components than the embodiment that has been described. Other types of components include “purchase” buttons, advertisements, feedback indicators (such as feedback indicator 336 in FIG. 3 d) and links to additional services and information. In addition, other embodiments of first graphic element 314 or second graphic element 326 may feature different sizes, shapes and appearances than the embodiment that has been described.
  • In this embodiment, the song currently being provided to the user is visually represented by the rightmost graphic element (i.e., second graphic element 326). After songs have been provided to the user, or otherwise discarded, the graphic elements corresponding to those songs are scrolled to the left across playlist panel 304 (in this example, approximately three graphic elements total can be visualized to the user). In the embodiment shown in FIG. 3 c, first graphic element 314 corresponds to a song that has already been provided to the user, while second graphic element 326 corresponds to a song that is currently being provided to the user. In one embodiment of the invention, the fact that second graphic element 326 is currently being provided to the user is emphasized by tinting, shading or otherwise de-emphasizing first graphic element 314, or highlighting, brightening or otherwise emphasizing second graphic element 326. In addition, playback bar 334 may be featured as a component of second graphic element 326 to indicate how much of the currently provided song has already been played. Of course, other embodiments may feature alternative ways of visually representing the playlist and/or the progression of the playlist, as well as fewer or more graphic elements and alternative ways for representing those graphic elements.
  • In the embodiment shown in FIG. 3 c, graphical user interface 208 also features volume control 340, playback controls 342, “Help” button 344, “Share” button 346, “Account” button 348 and “Guide” button 350. Volume control 340 adjusts the audible volume of content objects having audio that are provided to the user in accordance with embodiments of the invention. Playback controls 342 allow the user to pause or resume the playing of content objects. Playback controls 342 also allow the user to terminate playing of the current content object in favor of another content object.
  • The user's selecting of “Help” button 344 generates an on-screen pop-up menu providing clickable menu choices that provide additional features to the user and enhance the user-friendliness of graphical user interface 208. For example, the on-screen pop-up menu may include choices providing additional information about a music discovery service, such as a FAQ, contact information or legal notices.
  • The user's selecting of “Share” button 346 generates another pop-up menu providing clickable menu choices relating to, for example, sharing features of graphical music interface 208. For example, the pop-up menu may include choices for providing a playlist to other users of the music discovery service (e.g., enabling another user to enjoy a station such as the station corresponding to “Station 1button 308 and thus to be provided content corresponding to that station). The pop-up menu may also include choices for facilitating the providing of content by another station created by another user, the operator or a third party.
  • The user's selecting of “Account” button 348 generates another pop-up menu providing clickable menu choices relating to, for example, customized user preferences or settings. For example, the pop-up menu may include choices for viewing favorite stations, editing account and contact information or subscribing to the music discovery service. “Subscribing” may mean, for example, that in exchange for an annual fee, the user will no longer see advertisement 216 when using the music discovery service.
  • In this embodiment, the user's selecting of “Guide” button 350 generates another pop-up menu providing clickable menu choices relating to, for example, enabling the user to selectively provide feedback about a content object such as a song. In one embodiment, “Guide” button 350 serves as the primary interface for the “back-and-forth” conversation between the user and the music discovery service. For example, the pop-up menu may include choices for enabling the user to provide feedback corresponding to comments such as “I really like this song,” “I don't like this song,” or “I′m tired of this song.” This feedback can be used to customize, adapt and/or enhance the initial playlist generated in connection with a station so that it is more attuned to the preferences of the user.
  • As another example, the pop-up menu generated by selecting “Guide” button 350 may include other feedback options, such as reasons why the user likes or dislikes a certain song. Exemplary reasons that the user may select as reasons why he or she likes the song include “I like the artist,” “I like the song,” “I like the beat,” “I like the instrument being played,” “I like the meaning of the lyrics,” or “I like the genre.” Exemplary reasons that the user may select as reasons why he or she dislikes the song include “I don't like the artist,” “I don't like the vocals,” “I don't like the repetitiveness,” “The music is too ‘mainstream,’” or “The music is too loud.”
  • In response to feedback provided by the user, the playlist may be modified. Modifications to the playlist are accomplished, for example, by the use of weighing values and scaling functions as described in currently pending U.S. patent application Ser. No. 11/295,339, as will be discussed further below.
  • In addition, “Guide” button may include other choices that provide the user with information as to why a song is being played (i.e. what musicological attributes, such as characteristics or focus traits, are contained in a song). “Guide” button may also include other choices that enable the user to selectively modify the input seed so that it, and the playlist that is generated as a result of a comparison between the input seed and other songs, reflects additional artists or songs.
  • FIG. 4 depicts, in accordance with an embodiment of the invention, station pop-up menu 400, which is generated in response to a user selecting a button such as “Station 1button 308 in FIG. 3 c. Station pop-up menu 400 includes menu choices such as “Add More Music” menu choice 402, “Email This Station” menu choice 404, “Edit This Station” menu choice 406, “Rename This Station” menu choice 408 and “Delete This Station” menu choice 410. Other embodiments of the invention may have fewer, additional or alternative menu choices.
  • In one embodiment, “Add More Music” menu choice 402 enables the user to selectively modify the input seed corresponding to the current station. FIG. 5 depicts “Add More Music” panel 500 that appears on graphical user interface 208 after the user has selected “Add More Music” menu choice 402. “Add More Music” panel features entry field 502. Entry field 502 enables the user to selectively modify the input seed by entering, for example, another artist name or song name (in addition to the artist name, song name or other input seed previously entered) and then selecting “Add” button 504 (if the user does not desire to selectively modify the input seed, then the user selects “Close” button 506). The additional artist name or song name is then factored into the comparison between the input seed and songs contained in the “MUSIC GENOME PROJECT” database. One way to factor the additional artist name or song name into the comparison is to utilize confidence and weighting factors to assign, for example, more or less weight to the musicological attributes of the additional artist name or song name in view of the initial input seed. After the input seed has been selectively modified, “Add More Music” panel 500 disappears and graphical user interface 208 proceeds to present and provide content corresponding to the modified input seed in accordance with FIGS. 2 a and 2 b. In doing so, “Station 1button 308 may appear differently to reflect the modified input seed.
  • In one embodiment, “Email This Station” menu choice 404 enables the user to selectively provide a station, and thus a playlist, to another user. FIG. 6 depicts “Email This Station” panel 600 that appears on graphical user interface 208 after the user has selected “Email This Station” menu choice 404. “Email This Station” panel 600 features station field 602, email field 604 and message field 606. When selected by the user, station field 602 enables the user to select a station to selectively provide to another user. The stations available to selectively appear on a drop-down menu and may include stations created by the user, such as the station corresponding to “Station 1” button 308 (FIG. 3 c), or other stations.
  • Email field 604 enables the user to enter an email address corresponding to another user for which the user desires to selectively provide a station. Message field 606 enables the user to provide a message (such as regular text or HTML) to the user for which the station has been selectively provided.
  • After the user has entered information into station field 602, email field 604 and message field 606, the user selects “Share” button 608 to initiate the selective providing of a station to another user. The information is transmitted to playlist server 108 (FIG. 1). Playlist server 108 prepares an email including the information entered in message field 606 to the recipient user utilizing SMTP or other common protocols. The return address of the email corresponds to the email address provided by the user upon registration with the music discovery service. The email further includes a hypertext link to the URL of the music discovery service. The hypertext link includes a command line argument of an identifier corresponding to the station the user desires to selectively provide. If the recipient is already registered with the music discovery service, the station is automatically provided. If the recipient is not registered with the music discovery service, an anonymous registration is created and the hypertext link will direct the recipient to graphical user interface 208 as if the recipient were the anonymous registrant. If the user does not desire to selectively provide a station, and thus a playlist, to another user, then the user selects “Cancel” button 610.
  • In one embodiment, “Edit This Station” menu choice 406 enables the user to, among other things, selectively modify feedback about a content object such as a song. FIGS. 7 a-c depict “Edit This Station” panel 700 that appears after the user has selected “Edit This Station” menu choice 406. Turning to FIG. 7 a, “Edit This Station” panel 700 features station title 702, which displays the name of the station (such as the station corresponding to “Station 1” button 308) that is being edited. “Edit This Station” panel 700 also features “Items You Added” panel 704, “Songs You Liked” panel 706 and “Songs You Didn't Like” panel 708. The user may access each of these panels by selecting tab 710 that corresponds to the appropriate panel.
  • “Items You Added” panel 704 features song name text 712 and/or artist name text 714 corresponding to selective modifications of the input seed corresponding to the current station. Thus, for example, song name text 712 and artist name text 714 respectively correspond to a song and artist previously entered by the user in order to selectively modify the input seed. The user may remove, for example, a song that had previously selectively modified the input seed by selecting “Remove” button 716. Thereafter, graphical user interface 208 will no longer present and provide content corresponding to the modified input seed. Instead, graphical user interface 208 will proceed to present and provide content corresponding to, for example, the initial input seed, or to the input seed as selectively modified by entry of artist 714.
  • “Songs You Liked” panel 706 features, for example, song name text 718 (or artist name text) corresponding to selective feedback that the user has provided about a song. Thus, for example, song name text 712 corresponds to a song for which the user has previously selectively provided positive feedback. In addition, “Songs You Didn't Like” panel 708 features, for example, song name text 722 (or artist name text) corresponds to a song for which the user has previously selectively provided negative feedback.
  • The user may delete the feedback previously provided by selecting “Remove” button 720. Thereafter, when the song is provided, graphical user interface 208 will no longer display feedback indicator 336 (FIG. 3 d). Multiple songs and/or artists may be listed on “Items You Added” panel 704, “Songs You Liked” panel 706 or “Songs You Didn't Like” panel 708. Moreover, the feedback about the song will no longer be utilized in connection with generating playlists.
  • As stated previously, in one embodiment, station pop-up menu 400 also features “Rename This Station” menu choice 408 and “Delete This Station” menu choice 410. “Rename This Station” menu choice 408 enables the user to selectively provide an edited name for, for example, the station that corresponds to “Station 1button 308. “Delete This Station” menu choice 410 enables the user to remove a station from graphical interface 208.
  • FIG. 8 depicts, in accordance with an embodiment of the invention, content pop-up menu 800, which is generated in response to a user selecting a component of a graphical element, such as content art 332 of second graphic element 326. Content pop-up menu 800 includes menu choices such as “Why Did You Play This Song” menu choice 802, “Make a New Station from This Song” menu choice 804, “Buy This Song” menu choice 806 and “Buy This Album” menu choice 808. Other embodiments of the invention may have fewer, additional or alternative menu choices.
  • In one embodiment, “Why Did You Play This Song” menu choice 802 initiates the selectively providing of information to the user. FIG. 9 depicts “Information” panel 900 that appears on graphical user interface 208 after the user has selected, for example, “Why Did You Play This Song” menu choice 802. “Information” panel 900 features information, such as information provided in information text 902, about the song or other content object currently being provided to the user. For example, “Information” panel 900 may include information relating to a characteristic or focus trait of the song or other content object. Alternatively, “Information” panel 900 may also include information relating to background knowledge about the song, the artist who created the song or other relevant information. To the extent the user no longer desires to review the information, the user selects “Close” button 904 and information panel 900 disappears.
  • In one embodiment, “Make a New Station from This Song” menu choice 804 facilitates the presenting of content to a user in accordance with the present invention. FIG. 10 depicts “Create New Station” panel 1000 that appears on graphical user interface 208 after the user has selected, for example, “Make a New Station from This Song” menu choice 804. “Create New Station” panel 1000 features input seed field 1002 and “Create” button 1004. In one embodiment of the invention, input seed field 1002 is automatically filled with the song name corresponding to the song that was provided when content pop-up menu 800 was initially selected. In another embodiment, input seed field 1002 is empty and awaits the entry of a song name by the user. To initiate the creation of a new station, the user selects “Create” button 1004 after input seed field 1002 has been filled. In another embodiment, a station is automatically created in graphical user interface 208 after the user has selected “Make a New Station from This Song” menu choice 804. To the extent the user does not desire to create a new station, the user selects “Close” button 1006. “Create New Station” panel 1000 disappears and is replaced on the display by graphical user interface 208.
  • In one embodiment, content pop-up menu 800 features “Buy This Song” menu choice 806 and “Buy This Album” menu choice 808. If the user selects “Buy This Song” menu choice 806, then the selective purchase of the song (or other content object) is enabled. One way to enable the selective purchase of the song is to hyperlink “Buy This Song” menu choice 806 to a web site such as the “iTunes” web site from Apple Computer Corp. that offers songs for sale. The hyperlink may include a general URL as well as a parameter specifying the exact song for purchase. If the user selects “Buy This Album” menu choice 808, then the selective purchase of the album (or other content object) is enabled. One way to enable the selective purchase of the album is to hyperlink “Buy This Album” menu choice 808 to a web site such as the web site of Amazon.com, which sells albums. The hyperlink may include a general URL as well as a parameter specifying the exact song for purchase.
  • Content pop-up menu 800 also includes menu choices such as “I Like It” menu choice 810 and “I Don't Like It” menu choice 812. “I Like It” menu choice 810 and “I Don't Like It” menu choice 812 enable the user to selectively provide, respectively, positive or negative feedback about the current song or other content object. If the user selects “I Like It” menu choice 810, then feedback indicator 336 in the shape of, for example, a “thumbs-up” sign is displayed on graphic user interface 208 (FIG. 3 d). If the user selects “I Don't Like It” menu choice 812, then feedback indicator 336 in the shape of, for example, a “thumbs-down” sign is displayed on graphic user interface 208 (FIG. 3 d). Other types of feedback, such as “Don't play this song for awhile” may also be selectively provided. As stated previously, feedback may be used to customize and enhance playlists and other aspects of the user experience.
  • Content pop-up menu 800 further includes “Add to Favorites” menu choice 814. In one embodiment, “Add to Favorites” menu choice 814 enables the user to selectively associate the song or other content object with a favorites list. FIG. 11 depicts “Favorites” display 1100. “Favorites” display 1100 may appear, for example, as a panel in graphical user interface 208 or as a separate web page provided by playlist server 108. Another way for the user to access “Favorites” display 1100 is by selecting “Favorites” button 210 (FIG. 2 a). “Favorites” display 1100 keeps track of songs that the user has identified as good or otherwise significant. In one embodiment, “Favorites” display 1100 features management icons 1102 and 1104, song text 1106, artist text 1108 and station text 1110. Management icons 1102 and 1104 enable the user to remove and otherwise manipulate songs listed in the favorites list in “Favorites” display 1100. Song text 1106 and artist text 1108 provide information about the song that has been selectively associated with the “Favorites” list. Station text 1110 provides the name of the station, such as the station corresponding to the “Station 1button 308, from which the song was selectively associated with the “Favorites” list.
  • In one embodiment, “Favorites” display 1100 also features date 1112, album purchase icon 1114 and song purchase item 1116. Date 1112 provides information as to when the song was selectively associated with the “Favorites” list. Album purchase icon 1114 enables the selective purchase of the album (or other content object) from which the song originates. One way to enable the selective purchase of the album is to hyperlink album purchase icon 1114 to a web site such as the web site of Amazon.com, which sells albums. Song purchase icon 1116 enables the selective purchase of the song (or other content object). One way to enable the selective purchase of the song is to hyperlink song purchase icon 1116 to a web site such as the “iTunes” web site from Apple Computer Corp. that offers songs for sale.
  • It will be appreciated that the design of all displays, windows, interfaces, panels, graphic elements and other components discussed are not limited to the designs specified. Rather, such designs may be of any type or variety that is aesthetically pleasing or functional.
  • FIG. 12 depicts a flow diagram overview of a method for presenting and providing content to a user 1200 that can be executed in connection with, for example, the system depicted in FIG. 1.
  • In “Obtain Input Seed” step 1202 of FIG. 12, the user is enabled to selectively provide an input seed. As stated previously, the input seed may be a song name such as “Paint It Black” or even a group of songs such as “Paint It Black” and “Ruby Tuesday.” Alternatively, the input seed may be an artist name such as “Rolling Stones.” Other types of input seeds could include, for example, genre information such as “Classic Rock” or era information such as “1960s.” In one embodiment of the invention, the input seed is sent to playlist server 108 (FIG. 1) in order to perform the subsequent generation of a playlist. Encryption and other security methods may be used to protect communications between playlist server 108, content server 118 and/or terminals 102, 104 and 106.
  • In “Generate or Modify Playlist” step 1204, a playlist is first generated as a result of a comparison between the input seed and a plurality of database items. As stated previously, in one embodiment of the invention, the input seed is received from terminals 102, 104 and 106 and the playlist is generated on playlist server 108.
  • One or more embodiments of the invention utilize the “MUSIC GENOME PROJECT” database, which is a large database of records, each describing a single piece of music and an associated set of search and matching functions that operate on that database. The matching engine effectively calculates the distance between a source song and the other songs in the database and then sorts the results to yield an adjustable number of closest matches. Before continuing with FIG. 12, a method of generating or modifying a playlist will be discussed in accordance with one embodiment of the “MUSIC GENOME PROJECT” database will be discussed.
  • Song Matching
  • In the “MUSIC GENOME PROJECT” database, each song is described by a set of characteristics, or “genes”, or more that are collected into logical groups called “chromosomes.” The set of chromosomes make up the genome. One of these major groups in the genome is the “Music Analysis” Chromosome. This particular subset of the entire genome is sometimes referred to as “the genome.”
  • Each gene can be thought of as an orthogonal axis of a multi-dimensional space and each song as a point in that space. Songs that are geometrically close to one another are “good” musical matches. To maximize the effectiveness of the music matching engine, we maximize the effectiveness of this song distance calculation.
  • A given song “S” is represented by a vector containing approximately 150 genes. Each gene corresponds to a characteristic of the music, for example, gender of lead vocalist, level of distortion on the electric guitar, type of background vocals, etc. In a preferred embodiment, rock and pop songs have 150 genes, rap songs have 350, and jazz songs have approximately 400. Other genres of music, such as world and classical, have 300-500 genes. The system depends on a sufficient number of genes to render useful results. Each gene “s” of this vector has a value of an integer or half-integer between 0 and 5. However, the range of values for characteristics may vary and is not strictly limited to just integers or half-integers between 0 and 5.

  • Song S=s 2 , s 3 , s n)
  • The simple distance between any two songs “S” and “T”, in n-dimensional space, can be calculated as follows:

  • distance=square-root of (the sum over all n elements of the genome of (the square of (the difference between the corresponding elements of the two songs)))
  • This can be written symbolically as:

  • distance(S, T)=sqrt [(for i=1 to n) Σ(s i −t i)̂2]
  • Because the monotonic square-root function is used in calculating all of these distances, computing the function is not necessary. Instead, the invention uses distance-squared calculations in song comparisons. Accepting this and applying subscript notation, the distance calculation is written in simplified form as:

  • distance(S, T)=Σ(s−t)̂2
  • Weighted and Focus Matching
  • Weighted Matching
  • Because not all of the genes are equally important in establishing a good match, the distance is better calculated as a sum that is weighted according to each gene's individual significance. Taking this into account, the revised distance can be calculated as follows:

  • distance=Σ[w*(s−t)̂2]=[w 1*(s 1 −t 1)̂2]+[w 2*(s 2 −t 2)̂2]+ . . .
  • where the weighting vector “W,”

  • Song W=(w 1 , w 2 , w 3 , . . . w n)
  • is initially established through empirical work done, for example, by a music team that analyzes songs. The weighting vector can be manipulated in various ways that affect the overall behavior of the matching engine. This will be discussed in more detail later in this document.
  • Scaling Functions
  • The data represented by many of the individual genes is not linear. In other words, the distance between the values of 1 and 2 is not necessarily the same as the distance between the values of 4 and 5. The introduction of scaling functions f(x) may adjust for this non-linearity. Adding these scaling functions changes the matching function to read:

  • distance=Σ[w*(f(s)−f(t))̂2]
  • There are a virtually limitless number of scaling functions that can be applied to the gene values to achieve the desired result.
  • Alternatively, one can generalize the difference-squared function to any function that operates of the absolute difference of two gene values. The general distance function is:

  • distance=Σ[w*g(|s−t|)]
  • In the specific case, g(x) is simply x2, but it could become x3 for example if it was preferable to prioritize songs with many small differences over ones with a few large ones.
  • Focus Matching
  • Focus matching allows the end user of a system equipped with a matching engine to control the matching behavior of the system. Focus traits may be used to re-weight the song matching system and refine searches for matching songs to include or exclude the selected focus traits.
  • Focus Trait Presentation
  • Focus Traits are the distinguishing aspects of a song. When an end user enters a source song into the system, its genome is examined to determine which focus traits have been determined by music analysts to be present in the music. Triggering rules are applied to each of the possible focus traits to discover which apply to the song in question. These rules may trigger a focus trait when a given gene rises above a certain threshold, when a given gene is marked as a definer, or when a group of genes fits a specified set of criteria. The identified focus traits (or a subset) are presented on-screen to the user. This tells the user what elements of the selected song are significant.
  • Focus Trait Matching
  • An end user can choose to focus a match around any of the presented traits.
  • When a trait, or number of traits, is selected, the matching engine modifies its weighting vector to more tightly match the selection. This is done by increasing the weights of the genes that are specific to the Focus Trait selected and by changing the values of specific genes that are relevant to the Trait. The resulting songs will closely resemble the source song in the trait(s) selected.
  • Personalization
  • The weighting vector can also be manipulated for each end user of the system. By raising the weights of genes that are important to the individual and reducing the weights of those that are not, the matching process can be made to improve with each use.
  • Agreation
  • Song to Song Matching
  • The matching engine is capable of matching songs. That is, given a source song, it can find the set of songs that closely match it by calculating the distances to all known songs and then returning the nearest few. The distance between any two songs is calculated as the weighted Pythagorean sum of the squares of the differences between the corresponding genes of the songs.
  • Basic Multi-Song Matching
  • It may also be desirable to build functionality that will return the best matches to a group of source songs. Finding matches to a group of source songs is useful in a number of areas as this group can represent a number of different desirable searches. The source group could represent the collected works of a single artist, the songs on a given CD, the songs that a given end user likes, or analyzed songs that are known to be similar to an unanalyzed song of interest. Depending on the makeup of the group of songs, the match result has a different meaning to the end user but the underlying calculation should be the same.
  • This functionality provides a list of songs that are similar to the repertoire of an artist or CD. Finally, it will allow us to generate recommendations for an end user, purely on taste, without the need for a starting song.
  • FIG. 13 illustrates two songs. In this Figure, the song on the right is a better match to the set of source songs in the center.
  • Vector Pairs
  • Referring to FIG. 14, one way to implement the required calculation is to group the songs into a single virtual song that can represent the set of songs in calculations. The virtual “center” is defined to be a song vector whose genes are the arithmetic average of the songs in the original set. Associated with this center vector is a “deviation” vector that represents the distribution of the songs within the set. An individual gene that has a very narrow distribution of values around the average will have a strong affinity for the center value. A gene with a wide distribution, on the other hand, will have a weak affinity for the center value. The deviation vector will be used to modify the weighing vector used in song-to-song distance calculations. A small deviation around the center means a higher net weighting value.
  • The center-deviation vector pair can be used in place of the full set of songs for the purpose of calculating distances to other objects.
  • Raw Multi-Song Matching Calculation
  • If the assumption is made that a songs gene's are normally distributed and that they are of equal importance, the problem is straightforward. First a center vector is calculated and a standard deviation vector is calculated for the set of source songs. Then the standard song matching method is applied, but using the center vector in place of the source song and the inverse of the square of the standard deviation vector elements as the weights:

  • Target song vectors T=(t 2 , . . . t n)

  • Center vector of the source group C=(μ1, μ2, . . . , μn)

  • Standard deviation vector of the source group D=(σ1, σ2, . . . σn)

  • distancet=Σ[(1/σi)̂2*(μ1 −t i)̂2]
  • As is the case with simple song-to-song matching, the songs that are the smallest distances away are the best matches.
  • Using Multi-Song Matching With the Weighting Vector
  • The weighting vector that has been used in song-to-song matching must be incorporated into this system alongside the 1/σ̂2 terms. Assuming that they are multiplied together so that the new weight vector elements are simply:

  • New weight=w iî2
  • A problem that arises with this formula is that when σ2 is zero the new weight becomes infinitely large. Because there is some noise in the rated gene values, σ2 can be thought of as never truly being equal to zero. For this reason a minimum value is added to it in order to take this variation into account. The revised distance function becomes:

  • distancet=Σ[(w i*0.25/(σî2+0.25))*(μi t i)̂2]
  • Other weighting vectors may be appropriate for multi-song matching of this sort. Different multi-song weighting vector may be established, or the (0.5)2 constant may be modified to fit with empirically observed matching results.
  • Taste Portraits
  • Groups with a coherent, consistent set of tracks will have both a known center vector and a tightly defined deviation vector. This simple vector pair scheme will breakdown, however, when there are several centers of musical style within the collection. In this case we need to describe the set of songs as a set of two or more vector pairs.
  • As shown in FIG. 15, the song group can be described with two vector pairs. By matching songs to one OR the other of the vector pairs, we will be able to locate songs that fit well with the set. If we were to try to force all of the songs to be described by a single pair, we would return songs in the center of the large ellipse that would not be well matched to either cluster of songs.
  • Ideally there will be a small number of such clusters, each with a large number of closely packed elements. We can then choose to match to a single cluster at a time. In applications where we are permitted several matching results, we can choose to return a few from each cluster according to cluster size.
  • Returning to “Generate or Modify Playlist” step 1204 in FIG. 12, FIG. 16 shows a more detailed flow diagram for one or more embodiments of this step.
  • In “Identify Characteristics” step 1604 in FIG. 16, characteristics that correspond to the input seed are identified. As stated previously, characteristics may include, for example, gender of lead vocalist, level of distortion on the electric guitar, type of background vocals, etc. Characteristics may also include, for example, other types of musicological attributes such as syncopation, which is a shift of accent in a musical piece that occurs when a normally weak beat is stressed. In one or more embodiments of the invention, such characteristics are retrieved from one or more items corresponding to the input seed in a Music Genome Project database.
  • FIG. 17 shows a more detailed flow diagram for one embodiment of the
  • “Identify Characteristics” step 1604 (FIG. 16). As indicated previously, “Identify Characteristics” step 1604 as well as all of the other steps in FIG. 16, can be executed on, for example, the servers in FIG. 1.
  • In order to identify characteristics corresponding to the input seed, the input seed itself must first be analyzed as shown in “Input Seed Analysis” step 1702. Accordingly, database 112 in FIG. 1, which may be a Music Genome Project database, is accessed to first identify whether the input seed is an item in database 112. To the extent the input seed is not an item in the database, the user may be asked for more information in an attempt to determine, for example, whether the input seed was inputted wrong (e.g., “Beetles” instead of “Beatles”) or whether the input seed goes by another name in database (e.g., “I feel fine” instead of “She's in love with me”). Alternatively, close matches to the input seed may be retrieved from the database and displayed to the user for selection.
  • If the input seed is in the database, the input seed is then categorized. In the embodiment shown in FIG. 17, the input seed is categorized as either a “Song Name” or “Artist Name.” Such categorization is realized by, for example, retrieving “Song Name” or “Artist Name” information associated with the input seed from the database. Alternatively, such categorization is realized by asking the user whether the input seed is a “Song Name” or “Artist Name.”
  • If the input seed is a song name, then “Retrieve Characteristics” step 1704 is executed. In “Retrieve Characteristics” step 1704, a song vector “S” that corresponds to the song is retrieved from the database for later comparison to another song vector. As stated previously, in one embodiment the song vector contains approximately 150 characteristics, and may have 400 or more characteristics:

  • Song S=(s 1 , s 2 , s 3 , . . . , s n)
  • Each characteristic “s” of this vector has a value selected from a range of values established for that particular characteristic. For example, the value of the “syncopation” characteristic may be any integer or half-integer between 0 and 5. As an empirical example, the value of the syncopation characteristic for most “Pink Floyd” songs is 2 or 2.5. The range of values for characteristics may vary and is not limited to just integers or half-integers between 0 and 5.
  • If the input seed is an artist name, then (in the embodiment of FIG. 17) “Generate
  • Average” step 1706 is executed. In one embodiment of “Generate Average” step 1706, song vectors S1 to Sn, which each correspond to one of n songs in the database by the artist that is the subject of the input seed, are retrieved. Alternatively, and as stated previously, song vectors S1 to Sn could correspond to one of n songs in the database on a particular album by the artist.
  • After song vectors S1 to Sn have been retrieved, an average of all values for each characteristic of every song vector S1 to Sn is calculated and populated into a “center” or virtual song vector:

  • Center vector C=(μ1, μ2, . . . μn)

  • μ1=(s 1,1 +s 2,1 + . . . s n,1)/n
  • Of course, other statistical methods besides computing an average could be used to populate center vector “C.” Center vector “C” is then used for later comparison to another song vector as a representation of, for example, the average of all songs by the artist. In one embodiment of the invention, center vector “C1” corresponding to a first artist may be compared to center vector “C2” corresponding to a second artist.
  • After song vectors S1 to Sn have been retrieved, “assign confidence factor” step 1708 is executed. In “assign confidence factor” step 1708, a deviation vector “D” is calculated:

  • Deviation Vector D=(σ1, σ2, . . . σn)

  • σ=sprt(((s 1,1−μ1)̂2+(s 2,1μ1)̂2+(s n,1−μ1)̂2)/(n−1))
  • that shows how similar or dissimilar are the characteristics among each of song vectors S1 to Sn. While one embodiment of the invention contemplates populating the deviation vector by determining the standard deviation of all values for each characteristic of every song vector S1 to Sn, other statistical methods could also be used. As an empirical example of the use of standard deviation to calculate the deviation vector, the value of the syncopation characteristic for most “Pink Floyd” songs is 2 or 2.5, which results in a smaller standard deviation value (e.g., 0.035) than if a standard deviation value were calculated for a characteristic having more divergent values (e.g., if the value of the syncopation characteristic for all songs by Pink Floyd was more widely dispersed between 0 and 5).
  • To the extent a standard deviation value for a certain characteristic is larger, the averaged value of that characteristic in the virtual song vector is considered to be a less reliable indicator of similarity when the virtual song vector is compared to another song vector. Accordingly, as indicated previously, the values of the deviation vector serve as “confidence factors” that emphasize values in the virtual song vector depending on their respective reliabilities. One way to implement the confidence factor is by multiplying the result of a comparison between the center vector and another song vector by the inverse of the standard deviation value. Thus, for example, the confidence factor could have a value of 0.25/(σî2+0.25). The “0.25” is put into the equation to avoid a mathematically undefined result in the event σî2 is 0 (i.e., the confidence factor avoids “divide by zero” situations).
  • Returning to FIG. 16, “Identify Focus Traits” step 1606 identifies focus traits based on the values of characteristics of song vector (or virtual song vector) S. As stated previously, focus traits are specific combinations of characteristics (or even a single notable characteristic) representing significantly discernable attributes of a song. As such, focus traits are the kernel of what makes one song actually sound different, or like, another song. Focus traits may be created and defined in many ways, including by having trained musicologists determine what actually makes one song sound different from another, or else having users identify personal preferences (e.g., receiving input from a user stating that he/she likes songs with male lead vocals). Exemplary focus traits include “male lead vocal” or “Middle Eastern influence.” There can be 1, 10, 1000 or more than 1000 focus traits, depending on the desired complexity of the system.
  • In one embodiment of the invention, a set of rules known as “triggers” is applied to certain characteristics of song vector S to identify focus traits. For example, the trigger for the focus trait “male lead vocal” may require the characteristic “lead vocal present in song” to have a value of 5 on a scale of 0 to 5, and the characteristic “gender” to also have a value of 5 on a scale of 0 to 5 (where “0” is female and “5” is male). If both characteristic values are 5, then the “male lead vocal” focus trait is identified. This process is repeated for each focus trait. Thereafter, any identified focus traits may be presented to the user through the user interface.
  • Now that focus traits have been identified, “Weighting Factor Assignment” step 1608 is executed. In “weighting factor assignment” step 1608, comparative emphasis is placed on some or all of focus traits by assigning “weighting factors” to characteristics that triggered the focus traits. Alternatively, “weighting factors” could be applied directly to certain characteristics.
  • Accordingly, musicological attributes that actually make one song sound different from another are “weighted” such that a comparison with another song having those same or similar values of characteristics will produce a “closer” match. In one embodiment of the invention, weighting factors are assigned based on a focus trait weighting vector W, where w1, w2 and wn correspond to characteristics s1, s2 and sn of song vector S.

  • Weighting Vector W=(w 1 , w 2 , w 3 , . . . , w n)
  • In one embodiment of the invention, weighting vector W can be implemented into the comparison of songs having and song vectors “S” and “T” by the following formula:

  • distance (W, S, T)=Σw*(s−t)̂2
  • As described previously, one way to calculate weighting factors is through scaling functions. For example, assume as before that the trigger for the focus trait “male lead vocal” requires the characteristic “lead vocal present in song” to have a value of 5 on a scale of 0 to 5, and the characteristic “gender” to also have a value of 5 on a scale of 0 to 5 (where “0” is female and “5” is male).
  • Now assume the song “Yesterday” by the Beatles corresponds to song vector S and has an s1 value of 5 for the characteristic “lead vocal present in song” and an s2 value of 5 for the characteristic “gender.” According to the exemplary trigger rules discussed previously, “Yesterday” would trigger the focus trait “male lead vocal.” By contrast, assume the song “Respect” by Aretha Franklin corresponds to song vector T and has a t1 value of 5 for the characteristic “lead vocal present in song” and a t2 value of 0 for the characteristic “gender.” These values do not trigger the focus trait “male lead vocal” because the value of the characteristic “gender” is 0. Because a focus trait has been identified for characteristics corresponding to s1 and s2, weighting vector W is populated with weighting factors of, for example, 100 for w1 and w2. Alternatively, weighting vector W could receive different weighting factors for w1 and w2 (e.g., 10 and 1000, respectively).
  • In “Compare Weighted Characteristics” step 1610, the actual comparison of song vector (or center vector) S is made to another song vector T. Applying a comparison formula without a weighting factor, such as the formula distance(S, T)=(s−t)̂2, song vectors S and T would have a distance value of (s1−t1)̂2+(s2−t2)̂2, which is (5−5)̂2+(5−0)̂2, or 25. In one embodiment of the invention, a distance value of 25 indicates a close match.
  • By contrast, applying a comparison formula featuring weighting vector W produces a different result. Specifically, the weighting vector W may multiply every difference in characteristics that trigger a particular focus trait by 100. Accordingly the equation becomes w1(s1−t1)̂2+w2(s2−t2)̂2, which is 100(5−5)̂2+100(5−0)̂2, or 2500. The distance of 2500 is much further away than 25 and skews the result such that songs having a different gender of the lead vocalist are much less likely to match. By contrast, if song vector T corresponded to another song that did trigger the focus trait “male lead vocal” (e.g., it is “All I Want Is You” by U2), then the equation becomes 100(5−5)̂2+100(5−5)̂2, or 0, indicating a very close match.
  • As another example of one embodiment of the invention, a weighting vector value of 1,000,000 in this circumstance would effectively eviscerate any other unweighted matches of characteristics and means that, in most circumstances, two songs would never turn up as being similar.
  • As indicated previously, it is also possible for one or more values of the weighting vector to be assigned based on preferences of the user. Thus, for example, a user could identify a “male lead vocal” as being the single-most important aspect of songs that he/she prefers. In doing so, a weighting vector value of 10,000 may be applied to the comparison of the characteristics associated with the “male lead vocal” focus trait. As before, doing so in one embodiment of the invention will drown out other comparisons.
  • In one embodiment of the invention, one weighting vector is calculated for each focus trait identified in a song. For example, if 10 focus traits are identified in a song (e.g., “male lead vocalist” and 9 other focus traits), then 10 weighting vectors are calculated. Each of the 10 weighting vectors is stored for potential use during “Compare Weighted Characteristics” step 1610. In one embodiment of the invention, users can select which focus traits are important to them and only weighting vectors corresponding to those focus traits will be used during “Compare Weighted Characteristics” step 1610. Alternatively, weighting vectors themselves could be weighted to more precisely match songs and generate playlists.
  • In “Select Items” step 1612, the closest songs are selected for the playlist based on the comparison performed in “Compare Weighted Characteristics” step 1610. In one embodiment of the invention, the 20 “closest” songs are preliminary selected for the playlist and placed into a playlist set. Individual songs are then chosen for the playlist. One way to choose songs for the playlist is by random selection. For example, 3 of the 20 songs can be randomly chosen from the set. In one embodiment of the invention, another song by the same artist as the input seed is selected for the playlist before any other songs are chosen from the playlist. One way to do so is to limit the universe of songs in the database to only songs by a particular artist and then to execute the playlist generating method.
  • To the extent a set of weighted song vectors was obtained, a plurality of sets of closest songs are obtained. For example, if a song has 10 focus traits and the 20 closest songs are preliminarily selected for the playlist, then 10 different sets of 20 songs each (200 songs total) will be preliminarily selected. Songs can be selected for the playlist from each of the sets by, for example, random selection. Alternatively, each set can have songs be selected for the playlist in order corresponding to the significance of a particular focus trait.
  • As an alternative, or in addition to, randomly selecting songs for the playlist, rules may be implemented to govern the selection behavior. For example, aesthetic criteria may be established to prevent the same artist's songs from being played back-to-back after the first two songs, or to prevent song repetition within 4 hours.
  • Moreover, regulatory criteria may be established to comply with, for example, copyright license agreements (e.g., to prevent the same artist's songs from being played more than 4 times in 3 hours). To implement such criteria, a history of songs that have been played may be stored along with the time such songs were played.
  • Accordingly, songs are selected for the playlist from one or more playlist sets according to random selection, aesthetic criteria and/or regulatory criteria. To discern the actual order of songs in the playlist, focus traits can be ranked (e.g., start with all selected songs from the playlist set deriving from the “male lead vocal” focus trait and then move to the next focus trait). Alternatively, or in addition, the user can emphasize or de-emphasize particular playlist sets. If, for example, a user decides that he/she does not like songs having the focus trait of “male lead vocal,” songs in that playlist set can be limited in the playlist.
  • A number of songs are selected from the Set List and played in sequence as a Set. Selection is random, but limited to satisfy aesthetic and business interests, (e.g. play duration of a particular range of minutes, limits on the number of repetitions of a particular Song or performing artist within a time interval). A typical Set of music might consist of 3 to 5 Songs, playing for 10 to 20 minutes, with sets further limited such that there are no song repetitions within 4 hours and no more than 4 artist repetitions within 3 hours.
  • In one embodiment of the invention, the playlist features identifiers that correspond to, for example, song names. The identifiers may be index fields or other handles for content database 120 on content server 118. After the playlist has been generated, playlist server 108 may send an identifier corresponding to the input seed to the user at terminal 102, 104 or 106. To the extent the input seed was an artist name requiring the creation of a “center vector,” playlist server 108 may, for example, send an identifier corresponding to a song that is the closest match to the “center vector.” In one embodiment of the invention, a set of identifiers may be sent to terminal 102, 104 or 106 (or to multiple terminals) at once.
  • After an identifier is remotely provided to terminal 102, 104 or 106, the player on terminal 102, 104 or 106 proceeds to associate a graphic element (such as first graphic element 314 in FIG. 3) with the identifier. For example, content server 118 may store song name 316, artist name 318 and content art 320 in connection with a corresponding content object in content database 120. Accordingly, the player on terminal 102, 104 or 106 may request song name 316, artist name 318 and content art 320 that corresponds to the input seed or identifier from content server 118. Content server 118 then provides, in encrypted form, song name 316, artist name 318 and content art 320 to the player on terminal 102, 104 or 106.
  • In “Display Graphic Element” step 1208, first graphic element 314 appears in graphical user interface 208 as discussed previously. Song name 316, artist name 318 and content art 320 may be provided within first graphic element 314.
  • In “Provide Content Object” step 1210, a content object corresponding to the identifier or input seed is provided. For example, the player in terminal 102, 104 or 106 may send the identifier received from playlist server 108 to content server 118. In response, content server 118 may provide a content object corresponding to the identifier to the player on terminal 102, 104 or 106 and thus to the user.
  • Content server 118 may provide a content object to the user in several ways. For example, content server 118 may stream content object to the user through well-known streaming techniques and protocols such as User Datagram Protocol (UDP), Real Time Transport Protocol (RTP), Real Time Streaming Protocol (RTSP), Real Time Control Protocol (RTCP) and Transmission Control Protocol (TCP). As another example, content server 118 may provide a content object to the user through downloading. Thus, the content object is downloaded fully to terminal 102, 104 or 106 before it is provided to the user. As yet another example, the content object may be provided to the user through a hybrid method of streaming and downloading. In an embodiment of the invention, content server 118 may provide a content object at a rate of 10 to 20 times that of the playback rate. Portions of the content object that have not been played are cached in memory on terminal 102, 104 or 106.
  • In “Provide Content Object” step 1210, a content object corresponding to the identifier or input seed is provided. For example, the player in terminal 102, 104 or 106 may send the identifier received from playlist server 108 to content server 118. In response, content server 118 may provide a content object corresponding to the identifier to the player on terminal 102, 104 or 106 and thus to the user.
  • In “Obtain Feedback” step 1212, the user selectively provides feedback about a content object through graphical user interface 208 in the manner discussed previously. In one embodiment of the invention, feedback that has been selectively provided by the user is sent to playlist server 108. If the feedback about a content object is negative, then the playlist may be modified as discussed previously. For example, the user may selectively provide feedback that is negative about a song with a focus trait of “male lead vocal.” In response, a new playlist is generated by playlist server 108 (i.e., the existing playlist is modified) that accounts for the negative feedback. In one embodiment of the invention, a weighting value or scaling function corresponding to the focus trait of “male lead vocal” may be adjusted such that songs having strong focus traits of “male lead vocal” are less likely to match with the input seed originally provided by the user.
  • As another example, the user may selectively provide feedback that he or she does not like “jazz” music. “Jazz” may be a characteristic stored with regard to various songs in database 112. A weighting value of 1/1,000,000,000 is then assigned to the characteristic “jazz,” which means that a match between the input seed and “jazz” songs is unlikely to result from a comparison of the input seed and database items. Accordingly, the playlist will be modified to remove jazz songs.
  • Using Contextual Information to Generate and Modify Playlists
  • FIG. 18 shows a flow diagram of a method of generating a playlist of content objects in accordance with aspects of the invention. As explained above for “Obtain Feedback” step 1212 in FIG. 12, feedback about a content object may be obtained from a user and sent to playlist server 108. In “Store Feedback” step 1802, feedback from the user may be stored in memory on playlist server 108 (e.g., in database 112)(FIG. 1). Alternatively, the feedback may be stored in memory on a content server 118. One of skill in the art will appreciate that the feedback may be stored on any computing device containing memory that is accessible from network 110. For example, a separate feedback server and/or database may be beneficial in some configurations familiar to those of skill in the art.
  • The feedback stored in “Store Feedback” step 1802 may be positive, negative, or neutral. Numerous examples of negative feedback and positive feedback have been provided above. Meanwhile, in accordance with aspects of the invention, instances when the user refuses to give feedback or simply fails to give feedback are stored as neutral (or null) feedback. Neutral feedback tracks the number of occurrence of the absence of feedback. Therefore, in one aspect of the invention, the total number of times a particular song is played (i.e., “spins” of that song) is equal to the sum of the quantity of neutral feedbacks, positive feedbacks, and negative feedbacks stored for that song. One skilled in the art will appreciate that if the total number of spins, total number of positive feedback, and total number of negative feedback are stored for a song, then simple mathematics can be used to calculate the total number of neutral feedback. Therefore, the total number of each type of feedback need not necessarily be stored in accordance with aspects of the invention. At least one benefit of storing neutral feedback is that it may be useful in determining the weight to accord to feedback in the generation of playlists.
  • Furthermore, contextual information relating to the feedback is stored in “Store Contextual Information” step 1804. Similar to the storing of feedback, the contextual information may also be stored in memory on playlist server 108. Alternatively, the contextual information may be stored in memory on a content server 118. One of skill in the art will appreciate that the contextual information may be stored on any computing device containing memory that is accessible from network 110. For example, it may be beneficial in some configuration to setup a separate server and/or database for feedback and contextual information. FIG. 19, which is discussed in detail below, depicts an example of feedback and contextual information stored together in a database.
  • Each time a content object is presented to the user, feedback from the user (e.g., positive, negative, neutral) is stored along with the context (i.e., contextual information) in which the content object was presented. Examples of contextual information include, but are not limited to, the station on which the current song is playing, the song played immediately prior to the current song, the artist of the song playing immediately prior to the current song, one or more musicological characteristics of the prior song(s), the artist of the current song, the album containing the current song, time of day, day of week, calendar day (i.e., date), location of user (e.g., work, home, car, zip code, country, etc.), and user demographics (e.g., age, gender). Some contextual information may be related to the current song in the context of an extrinsic factors, such as location of user (i.e., geography), user demographics, and/or temporal (e.g., time of day, day of week, calendar date). Meanwhile, some contextual information may be related to the current song in the context of at least one of the following: particular artist, particular album, and/or particular station. In addition, some contextual information may be related to the current song in the context of a second content object (e.g., an immediately prior song). Moreover, some contextual information may be related to the current song in the context of characteristics of a second content object (e.g., characteristics of an immediately prior song such as treble, bass, vocals, etc.). Examples of each of the various types of contextual information will be clear to one of skill in the art after thorough review of the entirety of the disclosure.
  • Continuing with an earlier example where a user selectively provides feedback that is negative about a song (e.g., “With or Without You” by U2), contextual information relating to that feedback is automatically obtained and stored in memory. For example, the particular station (e.g., “Pet Shop Boys” Station) the user was listening to when the “With or Without You” song played, the particular artist (e.g., “Pet Shop Boys”) the user had provided as an input seed, and the time of day (e.g., 9 a.m.) and day of week (e.g., Monday) when the “With or Without You” song played may be automatically obtained and stored in memory as contextual information relating to the negative feedback. In another example, the fact that the user giving feedback was a 28-year old male located in Virginia may be stored as contextual information relating to the feedback.
  • In accordance with aspects of the invention, stored feedback and contextual information is retrieved in “Retrieve Feedback and Contextual Information” step 1806. In one example in accordance with aspects of the invention, when a user provides the artist/band “Pet Shop Boys” as an input seed, the stored feedback and contextual information related to the “Pet Shop Boys” Station is retrieved and used in generating a playlist of content objects (e.g., songs). In various embodiments in accordance with the invention, the feedback and contextual information are used in generating a playlist only after the quantity of feedback has exceeded a statistical threshold. In examples where the threshold has not been met, the generation of playlists may be similar to step 1204. In one example, the feedback and contextual information are used in playlist generation and/or optimization only after at least one hundred spins of a song in a particular context (e.g., the song “With or Without You” playing on “Pet Shop Boys” Station). In another example, the statistical threshold value may be set to 1,000 spins or 10,000 spins. One of skill in the art will appreciate that a larger threshold value may provide a higher level of confidence in the statistical reliability of the information collected, but may delay the resulting playlist optimizations.
  • Once the statistical threshold has been met, in “Generate Playlist Using Feedback and Contextual Information” step 1808, a playlist is generated by playlist server 108 that accounts for negative, positive, and neutral feedback provided in that particular context. Unlike the generation and modification of a playlist in step 1204, this step uses feedback with contextual information to deliver an enhanced playlist. In one example, the context in which a song is played may be a station having an input seed of artist “U2” (i.e., in the context of songs played on the “U2” Station). The feedback and contextual information stored for that particular context is depicted by the appropriate portion of database table 1900 depicted in FIG. 19.
  • Database table 1900 in FIG. 19 stores contextual information and feedback provided by users in the context of stations. In this example, for practical purposes, database table 1900 shows only some of the entries stored for a “U2” Station (i.e., a station based on an artist input seed) and a “Sweet Child of Mine” Station (i.e., a station based on a song input seed). Database table 1900 contains ten entries for each station, where each entry corresponds to a song that was played in the context of the “Sweet Child of Mine” Station or “U2” Station. In one example, the number of entries in database table 1900 may dynamically grow as different songs are played on that station and added to database table 1900. Alternatively, database table 1900 may be prepopulated with an entry corresponding to every available song. One skilled in the art will recognize that prepopulating database table 1900 may undesirably result in extremely large database tables in memory.
  • By way of example, database entry 1902 corresponds to the feedback provided by users listening to the “U2” Station in response to hearing artist U2's song “With or Without You.” In database table 1900, the “Artist Name” column contains “U2” and the “Song Name” column contains “With or Without You”. Other relevant portions of database table 1900 for entry 1902 are as follows: the “Spins” column contains 31,271 (i.e., the total number of times the song “With or Without You” has played on a “U2” Station), the “Positive” column contains the number of times a user gave positive feedback (e.g., by selecting “I Like It” menu choice 810 (FIG. 8) for the song) for the song, and the “Negative” column contains the number of times a user gave positive feedback (e.g., by selecting “I Don't Like It” menu choice 812 (FIG. 8) for the song) for the song. Meanwhile, as explained earlier, the quantity of neutral feedback can be obtained by simple arithmetic (i.e., Spins minus Positive minus Negative). In addition, the “Feedback” column is a total of positive and negative feedback. One skilled in the art will recognize that the “Feedback” column may be calculated using other columns in database table 1900 and need not be present in all embodiments.
  • In addition, database entry 1902 contains a “Song Q” column that stores values indicating the suitability of its respective content object in that particular context. In one example, the suitability quotient “Q” may be expressed as a numerical value between negative one and positive one. Negative “Q” values may be used to indicate that a particular content object (e.g., the song “With or Without You”) is less suitable in the particular context (e.g., on “Sweet Child of Mine” Station that song receives a negative 0.064 “Q” value). Meanwhile, a positive “Q” value may be used to indicate that a particular content object (e.g., the song “With or Without You”) is more suitable in the particular context (e.g., on “U2” Station that song receives a positive 0.045 “Q” value). In one embodiment, the “Q” value may be recalculated and updated in database table 1900 each time a song is played in the context of a particular station. One skilled in the art will appreciate that this may result in a great computational load on the relevant computing machines. Alternatively, feedback and contextual information may be collected and the corresponding “Q” values recalculated and updated only at regular intervals (e.g., 30 minutes or after 100 new spins).
  • The suitability quotient “Q” for a content object in a particular context is calculated using the function:

  • Q=((ΣF pos)−(ΣF neg))/(total # of spins)
  • where Fpos is the sum of all positive feedback for the content object, Fneg is the sum of all negative feedback for the content object, and “total # of spins” is the total number of times that a content object was performed in that particular context. For example, database entry 1902 shows that the song “With or Without You” by U2 was played on the “U2” Station 31,271 times. Of those times, the song received positive feedback 1,500 times and negative feedback 96 times. Using the suitability quotient function above, the “Q” value for the song “With or Without You” in the context of the “U2” Station is 0.045 (i.e., 1,500 minus 96, divided by 31,271). Therefore, a positive “Q” value for that song in that context will increase the likelihood that it will be found in the playlist generated in step 1808.
  • In one example, a weighting value or scaling function as discussed previously and that is appropriate for the suitability quotient “Q” value may be used to optimize the comparison performed by the matching algorithm. The matching algorithm was discussed extensively above. For example, according to the illustrative data in database table 1900 under the “U2” Station context, the song “Age of Consent” by New Order has a more favorable suitability quotient “Q” value than the song “We Are One” and is thus more likely to be selected for the playlist during a comparison.
  • In accordance with aspects of the invention, a number of content objects (e.g., songs) may be selected and played in a sequence determined, at least in part, using each song's respective “Q” value in the particular context. For example, according to the illustrative data in database table 1900 under the “U2” Station context, the song “Age of Consent” by New Order has a more favorable suitability quotient “Q” value than the song “We Are One” and is thus more likely to be selected for the playlist during a comparison.
  • In another example, the “Q” may be used to optimize the generation of a pleasing sequence of music (e.g., content objects) in a particular context. In accordance with aspects of the invention, a number of content objects (e.g., songs) may be selected and played in a sequence determined, at least in part, using each song's respective “Q” value in the particular context. For example, according to the illustrative data in database table 1900 under the “U2” Station context, the song “Age of Consent” by New Order has a more favorable suitability quotient “Q” value than the song “We Are One” and is thus more likely to be selected from a playlist to be played.
  • FIG. 20 depicts another flow diagram of a method of generating a playlist of content objects in accordance with aspects of the invention. In step 2002, an input seed is compared to a first database item corresponding to a first content object (i.e., song). The comparison may use the matching algorithm discussed above to identify a playlist of one or more songs that sound alike (e.g., are closely related to a source song or group of songs based on their characteristics and weighted comparisons of these characteristics). For example, a user may enter the artist “Massive Attack” as an input seed and be presented with one or more database items corresponding to content objects (e.g., the song “With or Without You”). In some examples a database item may be simply a unique identifier for a particular song.
  • In step 2004, contextual information relating to a first content object (e.g., the song “With or Without You”) is retrieved from memory. Many examples of contextual information were provided above. In addition, the suitability quotient “Q” calculated for the song “With or Without You” on the “Massive Attack” Station may be retrieved. The retrieving performed in step 2004 is similar to the retrieving performed in step 1806. As discussed earlier, the suitability quotient “Q” may optimize the actual order and/or placement of the songs in the playlist.
  • In step 2006, the contextual information relating to the first content object (e.g., the song “With or Without You”) is used to determine whether to select that object for the playlist. Step 2006, which is an enhancement of step 1612 (FIG. 16), uses retrieved contextual information in addition to the comparison algorithm used in step 1612 to select content objects for the playlist. For example, step 2006 may show that the song “With or Without You” corresponds to a negative “Q” value when played after the song “Angel”. Such information may be stored in a table in memory like table 1900 in FIG. 19. However, in this example the context in which the feedback was obtained would relate to the prior content object (e.g., the song “Angel”) and not the station, as was the case in a prior example.
  • The invention has been described with respect to specific examples including presently preferred modes of carrying out the invention. Those skilled in the art will appreciate that there are numerous variations and permutations of the above described systems and techniques, for example, that would be used with videos, wine, films, books and video games, that fall within the spirit and scope of the invention as set forth in the appended claims. The use of ordinal syntax such as “first” and “second” in this and related applications is intended for comprehension and readability purposes and is not intended to be limiting in the specification or in and/or between claims. For example, a first content object does not necessarily require the content object to come before a second content object.

Claims (36)

1-20. (canceled)
21. A computer implemented method comprising:
modeling each of a plurality of songs in a database as a vector corresponding to n dimensions of the song and storing each vector to form a vector database for each of the plurality of songs;
receiving a selected song and obtaining the vector that corresponds to the selected song from the vector database;
comparing the selected song vector to another database song vector; and
selecting at least one match song based on the comparison between the selected song vector and the database song vector.
22. The computer implemented method of claim 21 wherein a comparison value is calculated based on the selected song vector and the database song vector and the comparison value is utilized to determine the proximity of the selected song and the database song.
23. The computer implemented method of claim 21 further comprising calculating the distance between the selected song database vector and the database song vector.
24. The computer implemented method of claim 21 further comprising comparing the selected song vector to multiple vectors in the vector database to select a group of songs.
25. The computer implemented method of claim 24 further comprising creating a playlist from the group of songs.
26. The computer implemented method of claim 25 further comprising providing an option to the user to prevent the same artist's songs from being played back-to-back in the playlist.
27. The computer implemented method of claim 21 further comprising obtaining song matches to a group of selected songs by comparing the selected song vectors to multiple song vectors from the vector database.
28. The computer implemented method of claim 21 wherein a graphical user interface provides the user with an option to provide feedback input about the at least one match song.
29. The computer implemented method of claim 25 wherein a graphical user interface provides the user with an option to provide feedback input about at least one of the group of songs in the playlist and modifying the playlist based on the feedback input.
30. The computer implemented method of claim 21 wherein a plurality of songs in the song database are ranked based on comparing a plurality of song vectors from the vector database with the selected song.
31. A computer implemented method comprising:
representing each of a plurality of items in a database as a vector and storing each vector to form a vector database;
obtaining the vector that corresponds to a selected item from the vector database;
comparing the selected item vector to another database item vector; and
selecting at least one match item based on the comparison between the selected item vector and the database item vector.
32. The computer implemented method of claim 31 wherein a comparison value is calculated based on the selected item vector and the database item vector and the comparison value is utilized to determine the proximity of the selected item and the database item.
33. The computer implemented method of claim 31 further comprising calculating the distance between the selected item database vector and the database item vector.
34. The computer implemented method of claim 31 further comprising comparing the selected item vector to multiple vectors in the vector database to select a group of items.
35. The computer implemented method of claim 31 further comprising obtaining item matches to a group of selected items by comparing the selected item vectors to multiple item vectors from the vector database.
36. The computer implemented method of claim 31 wherein a graphical user interface provides the user with an option to provide feedback input about the at least one match item.
37. The computer implemented method of claim 31 wherein a plurality of items are ranked based on comparing a plurality of item vectors from the vector database with the source item.
38. A computer implemented method comprising:
modeling each of a plurality of songs as a vector and storing each vector to form a vector database;
receiving an input seed and obtaining an input seed vector;
comparing the input seed vector to another database song vector; and
selecting at least one match song based on the comparison between the input seed vector and the database song vector.
39. The computer implemented method of claim 38 further comprising determining whether the input seed is a song or an artist.
40. The computer implemented method of claim 39 wherein if the input seed is a song analyzing whether the input seed is in the database and obtaining the corresponding vector from the vector database.
41. The computer implemented method of claim 39 wherein if the input seed is an artist generating an average vector based on songs by that artist.
42. The computer implemented method of claim 38 wherein a comparison value is calculated based on the input seed vector and the database song vector and the comparison value is utilized to determine the proximity of the input seed and the database song.
43. The computer implemented method of claim 38 further comprising calculating the distance between the input seed vector and the database song vector.
44. The computer implemented method of claim 38 further comprising comparing the input seed vector to multiple vectors in the vector database to select a group of songs.
45. The computer implemented method of claim 44 further comprising creating a playlist from the group of songs.
46. The computer implemented method of claim 38 wherein a graphical user interface provides the user with an option to provide feedback input about the at least one match song.
47. The computer implemented method of claim 45 wherein a graphical user interface provides the user with an option to provide feedback input about at least one of the group of songs in the playlist and modifying the playlist based on the feedback input.
48. The computer implemented method of claim 38 wherein a plurality of songs in the song database are ranked based on comparing a plurality of song vectors from the vector database with the input seed song.
49. An apparatus comprising: a processor; and a memory for storing computer readable instructions that, when executed by the processor, cause the computer to perform the steps of:
creating a database comprising multiple items;
representing each item in the database as a vector and storing each vector to form a vector database;
obtaining the vector that corresponds to a selected item from the vector database;
comparing the selected item vector to another database item vector; and
selecting at least one match item based on the comparison between the selected item vector and the database item vector.
50. The apparatus of claim 49 wherein a comparison value is calculated based on the selected item vector and the database item vector and the comparison value is utilized to determine the proximity of the selected item and the database item.
51. The apparatus of claim 49 further comprising calculating the distance between the selected item database vector and the database item vector.
52. The apparatus of claim 49 further comprising comparing the selected item vector to multiple vectors in the vector database to select a group of items.
53. The apparatus of claim 49 further comprising obtaining item matches to a group of selected items by comparing the selected item vectors to multiple item vectors from the vector database.
54. The apparatus of claim 49 wherein a graphical user interface provides the user with an option to provide feedback input about the at least one match item.
55. The apparatus of claim 49 wherein a plurality of items in the item database are ranked based on comparing a plurality of item vectors from the vector database with the selected item.
US13/669,412 2001-05-16 2012-11-05 Methods and Systems for Utilizing Contextual Feedback to Generate and Modify Playlists Pending US20130179439A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/669,412 US20130179439A1 (en) 2001-05-16 2012-11-05 Methods and Systems for Utilizing Contextual Feedback to Generate and Modify Playlists

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
US29182101P 2001-05-16 2001-05-16
US10/150,876 US7003515B1 (en) 2001-05-16 2002-05-16 Consumer item matching method and system
US11/295,339 US20060206478A1 (en) 2001-05-16 2005-12-06 Playlist generating methods
US11/279,567 US20060212442A1 (en) 2001-05-16 2006-04-13 Methods of Presenting and Providing Content to a User
US11/380,563 US7962482B2 (en) 2001-05-16 2006-04-27 Methods and systems for utilizing contextual feedback to generate and modify playlists
US13/108,184 US8306976B2 (en) 2001-05-16 2011-05-16 Methods and systems for utilizing contextual feedback to generate and modify playlists
US13/669,412 US20130179439A1 (en) 2001-05-16 2012-11-05 Methods and Systems for Utilizing Contextual Feedback to Generate and Modify Playlists

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/108,184 Continuation US8306976B2 (en) 2001-05-16 2011-05-16 Methods and systems for utilizing contextual feedback to generate and modify playlists

Publications (1)

Publication Number Publication Date
US20130179439A1 true US20130179439A1 (en) 2013-07-11

Family

ID=37806244

Family Applications (3)

Application Number Title Priority Date Filing Date
US11/380,563 Expired - Lifetime US7962482B2 (en) 2001-05-16 2006-04-27 Methods and systems for utilizing contextual feedback to generate and modify playlists
US13/108,184 Expired - Lifetime US8306976B2 (en) 2001-05-16 2011-05-16 Methods and systems for utilizing contextual feedback to generate and modify playlists
US13/669,412 Pending US20130179439A1 (en) 2001-05-16 2012-11-05 Methods and Systems for Utilizing Contextual Feedback to Generate and Modify Playlists

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US11/380,563 Expired - Lifetime US7962482B2 (en) 2001-05-16 2006-04-27 Methods and systems for utilizing contextual feedback to generate and modify playlists
US13/108,184 Expired - Lifetime US8306976B2 (en) 2001-05-16 2011-05-16 Methods and systems for utilizing contextual feedback to generate and modify playlists

Country Status (2)

Country Link
US (3) US7962482B2 (en)
WO (1) WO2007067250A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090077052A1 (en) * 2006-06-21 2009-03-19 Concert Technology Corporation Historical media recommendation service
US20120066393A1 (en) * 2010-09-15 2012-03-15 Core Mobile Networks, Inc. System and method for real time delivery of context based content from the cloud to mobile devices
WO2015094558A1 (en) * 2013-12-19 2015-06-25 Gracenote, Inc. Media service
WO2015120068A1 (en) * 2014-02-04 2015-08-13 Google Inc. Adaptive music and video recommendations
US20160196270A1 (en) * 2013-12-19 2016-07-07 Gracenote, Inc. Station library creaton for a media service
US9471572B1 (en) 2013-12-31 2016-10-18 Google Inc. Recommending candidates for consumption
US20170124187A1 (en) * 2014-04-17 2017-05-04 Renault S.A.S. Method of adaptation of a representation of musical content
US11032290B2 (en) 2010-09-15 2021-06-08 Core Mobile Networks, Inc. Context-based analytics and intelligence

Families Citing this family (218)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020002039A1 (en) 1998-06-12 2002-01-03 Safi Qureshey Network-enabled audio device
US7068729B2 (en) 2001-12-21 2006-06-27 Digital Fountain, Inc. Multi-stage code generator and decoder for communication systems
US6307487B1 (en) 1998-09-23 2001-10-23 Digital Fountain, Inc. Information additive code generator and decoder for communication systems
WO2001035667A1 (en) * 1999-11-10 2001-05-17 Launch Media, Inc. Internet radio and broadcast method
US6389467B1 (en) 2000-01-24 2002-05-14 Friskit, Inc. Streaming media search and continuous playback system of media resources located by multiple network addresses
US7206775B2 (en) * 2000-07-06 2007-04-17 Microsoft Corporation System and methods for the automatic transmission of new, high affinity media
WO2002005140A1 (en) * 2000-07-11 2002-01-17 Launch Media, Inc. Online playback system with community bias
US7035873B2 (en) 2001-08-20 2006-04-25 Microsoft Corporation System and methods for providing adaptive media property classification
US9240810B2 (en) 2002-06-11 2016-01-19 Digital Fountain, Inc. Systems and processes for decoding chain reaction codes through inactivation
JP4039158B2 (en) 2002-07-22 2008-01-30 ソニー株式会社 Information processing apparatus and method, information processing system, recording medium, and program
CN100539439C (en) 2002-10-05 2009-09-09 数字方敦股份有限公司 The system coding of chain reaction sign indicating number and decode system and method
JP4241167B2 (en) * 2003-04-24 2009-03-18 ソニー株式会社 Program, data processing method, and data processing apparatus
KR101170629B1 (en) * 2003-10-06 2012-08-02 디지털 파운튼, 인크. Error-correcting multi-stage code generator and decoder for communication systems having single transmitters or multiple transmitters
JP4492190B2 (en) * 2004-04-07 2010-06-30 ソニー株式会社 Information processing apparatus and method, program
US9826046B2 (en) 2004-05-05 2017-11-21 Black Hills Media, Llc Device discovery for digital entertainment network
US8028038B2 (en) 2004-05-05 2011-09-27 Dryden Enterprises, Llc Obtaining a playlist based on user profile matching
US8028323B2 (en) 2004-05-05 2011-09-27 Dryden Enterprises, Llc Method and system for employing a first device to direct a networked audio device to obtain a media item
KR101205758B1 (en) 2004-05-07 2012-12-03 디지털 파운튼, 인크. File download and streaming system
WO2006020826A2 (en) * 2004-08-11 2006-02-23 Digital Fountain, Inc. Method and apparatus for fast encoding of data symbols according to half-weight codes
US7373110B1 (en) * 2004-12-09 2008-05-13 Mcclain John Personal communication system, device and method
US20060184462A1 (en) * 2004-12-10 2006-08-17 Hawkins Jeffrey C Methods, architecture, and apparatus for implementing machine intelligence and hierarchical memory systems
US7739208B2 (en) * 2005-06-06 2010-06-15 Numenta, Inc. Trainable hierarchical memory system and method
JP4835053B2 (en) * 2005-07-05 2011-12-14 ソニー株式会社 Content reproduction system, content provision method, content reproduction apparatus, content provision apparatus, content reproduction program, and content provision program
US7668821B1 (en) 2005-11-17 2010-02-23 Amazon Technologies, Inc. Recommendations based on item tagging activities of users
US7941389B2 (en) * 2006-02-10 2011-05-10 Numenta, Inc. Hierarchical temporal memory based system including nodes with input or output variables of disparate properties
US20070192267A1 (en) 2006-02-10 2007-08-16 Numenta, Inc. Architecture of a hierarchical temporal memory based system
US8732098B2 (en) 2006-02-10 2014-05-20 Numenta, Inc. Hierarchical temporal memory (HTM) system deployed as web service
US20080208966A1 (en) * 2007-02-28 2008-08-28 Numenta, Inc. Hierarchical Temporal Memory (HTM) System Deployed as Web Service
CN101686107B (en) 2006-02-13 2014-08-13 数字方敦股份有限公司 Streaming and buffering using variable FEC overhead and protection periods
US9270414B2 (en) 2006-02-21 2016-02-23 Digital Fountain, Inc. Multiple-field based code generator and decoder for communications systems
EP1826716A1 (en) * 2006-02-22 2007-08-29 Sony Deutschland Gmbh Method for updating a user profile
JP2007294036A (en) * 2006-04-26 2007-11-08 Sony Corp Information processing device and method, and program
US7971129B2 (en) 2006-05-10 2011-06-28 Digital Fountain, Inc. Code generator and decoder for communications systems operating using hybrid codes to allow for multiple efficient users of the communications systems
US9178535B2 (en) 2006-06-09 2015-11-03 Digital Fountain, Inc. Dynamic stream interleaving and sub-stream based delivery
US9380096B2 (en) 2006-06-09 2016-06-28 Qualcomm Incorporated Enhanced block-request streaming system for handling low-latency streaming
US9432433B2 (en) 2006-06-09 2016-08-30 Qualcomm Incorporated Enhanced block-request streaming system using signaling or block creation
US9386064B2 (en) * 2006-06-09 2016-07-05 Qualcomm Incorporated Enhanced block-request streaming using URL templates and construction rules
US9209934B2 (en) 2006-06-09 2015-12-08 Qualcomm Incorporated Enhanced block-request streaming using cooperative parallel HTTP and forward error correction
US9419749B2 (en) 2009-08-19 2016-08-16 Qualcomm Incorporated Methods and apparatus employing FEC codes with permanent inactivation of symbols for encoding and decoding processes
US7831928B1 (en) 2006-06-22 2010-11-09 Digg, Inc. Content visualization
US9219767B2 (en) * 2006-06-22 2015-12-22 Linkedin Corporation Recording and indicating preferences
US9015334B2 (en) * 2006-07-26 2015-04-21 Harris Technology, Llc Media player with automatic streaming of media files
US7521620B2 (en) * 2006-07-31 2009-04-21 Hewlett-Packard Development Company, L.P. Method of and system for browsing of music
US8775237B2 (en) * 2006-08-02 2014-07-08 Opinionlab, Inc. System and method for measuring and reporting user reactions to advertisements on a web page
US9202509B2 (en) 2006-09-12 2015-12-01 Sonos, Inc. Controlling and grouping in a multi-zone media system
US8483853B1 (en) 2006-09-12 2013-07-09 Sonos, Inc. Controlling and manipulating groupings in a multi-zone media system
US8788080B1 (en) 2006-09-12 2014-07-22 Sonos, Inc. Multi-channel pairing in a media system
US7827054B2 (en) * 2006-09-29 2010-11-02 Ourstage, Inc. Online entertainment network for user-contributed content
US9008634B2 (en) * 2006-10-06 2015-04-14 Napo Enterprises, Llc System and method for providing media content selections
CA2667573C (en) 2006-10-24 2018-10-23 Slacker, Inc. Method and device for playback of digital media content
US20100106852A1 (en) * 2007-10-24 2010-04-29 Kindig Bradley D Systems and methods for providing user personalized media content on a portable device
US8712563B2 (en) * 2006-10-24 2014-04-29 Slacker, Inc. Method and apparatus for interactive distribution of digital content
US10657168B2 (en) 2006-10-24 2020-05-19 Slacker, Inc. Methods and systems for personalized rendering of digital media content
US20080097967A1 (en) * 2006-10-24 2008-04-24 Broadband Instruments Corporation Method and apparatus for interactive distribution of digital content
WO2008067326A2 (en) * 2006-11-28 2008-06-05 Numenta, Inc. Group-based temporal pooling
US20080156173A1 (en) * 2006-12-29 2008-07-03 Harman International Industries, Inc. Vehicle infotainment system with personalized content
US9208251B2 (en) * 2007-01-19 2015-12-08 Linkedin Corporation Computer-based evaluation tool for selecting personalized content for users
DE102007006432B4 (en) * 2007-02-05 2010-07-08 Arndt-Helge Grap Apparatus and method for providing data
US20080195239A1 (en) * 2007-02-08 2008-08-14 Ben Aaron Rotholtz Collaborative playlist system and method
US20080261512A1 (en) * 2007-02-15 2008-10-23 Slacker, Inc. Systems and methods for satellite augmented wireless communication networks
US8037010B2 (en) * 2007-02-28 2011-10-11 Numenta, Inc. Spatio-temporal learning algorithms in hierarchical temporal networks
US7941392B2 (en) * 2007-02-28 2011-05-10 Numenta, Inc. Scheduling system and method in a hierarchical temporal memory based system
WO2008106624A2 (en) * 2007-02-28 2008-09-04 Slacker, Inc. Antenna array for a hi/lo antenna beam pattern and method of utilization
US8112367B2 (en) * 2007-02-28 2012-02-07 Numenta, Inc. Episodic memory with a hierarchical temporal memory based system
CA2680281C (en) * 2007-03-08 2019-07-09 Slacker, Inc. System and method for personalizing playback content through interaction with a playback device
US20080305736A1 (en) * 2007-03-14 2008-12-11 Slacker, Inc. Systems and methods of utilizing multiple satellite transponders for data distribution
CA2680797C (en) * 2007-03-14 2018-02-13 Slacker, Inc. Systems and methods for portable personalized radio
US20080235588A1 (en) * 2007-03-20 2008-09-25 Yahoo! Inc. Media player playlist creation and editing within a browser interpretable document
EP1975813A1 (en) 2007-03-31 2008-10-01 Sony Deutschland Gmbh Method for content recommendation
US8200663B2 (en) * 2007-04-25 2012-06-12 Chacha Search, Inc. Method and system for improvement of relevance of search results
JP4646249B2 (en) * 2007-06-06 2011-03-09 株式会社スクウェア・エニックス Program recording medium, portable video game machine, playback control program, and playback control method
US9918218B2 (en) 2007-06-12 2018-03-13 Avago Technologies General Ip (Singapore) Pte. Ltd. Method and system for a networked self-configuring communication device utilizing user preference information
US8260787B2 (en) * 2007-06-29 2012-09-04 Amazon Technologies, Inc. Recommendation system with multiple integrated recommenders
US8751507B2 (en) * 2007-06-29 2014-06-10 Amazon Technologies, Inc. Recommendation system with multiple integrated recommenders
US7949659B2 (en) * 2007-06-29 2011-05-24 Amazon Technologies, Inc. Recommendation system with multiple integrated recommenders
US8156118B2 (en) * 2007-08-20 2012-04-10 Samsung Electronics Co., Ltd. Method and system for generating playlists for content items
US8239877B2 (en) * 2007-08-28 2012-08-07 International Business Machines Corporation Pre-population of meta data cache for resolution of data marshaling issues
US9218415B2 (en) 2007-08-29 2015-12-22 Iheartmedia Management Services, Inc. System and method for providing broadcast listener participation
KR101129260B1 (en) 2007-09-12 2012-03-27 디지털 파운튼, 인크. Generating and communicating source identification information to enable reliable communications
US8285761B2 (en) * 2007-10-26 2012-10-09 Microsoft Corporation Aggregation of metadata associated with digital media files
US8645844B1 (en) 2007-11-02 2014-02-04 Ourstage, Inc. Comparison selection, ranking, and anti-cheating methods in an online contest environment
US10600082B1 (en) 2007-12-05 2020-03-24 Beats Music, Llc Advertising selection
US8175985B2 (en) * 2008-03-19 2012-05-08 Numenta, Inc. Plugin infrastructure for hierarchical temporal memory (HTM) system
US8554891B2 (en) * 2008-03-20 2013-10-08 Sony Corporation Method and apparatus for providing feedback regarding digital content within a social network
US7983998B2 (en) * 2008-03-21 2011-07-19 Numenta, Inc. Feedback in group based hierarchical temporal memory system
US8725740B2 (en) * 2008-03-24 2014-05-13 Napo Enterprises, Llc Active playlist having dynamic media item groups
UA99649C2 (en) 2008-04-07 2012-09-10 Косс Корпорейшн Normal;heading 1;WIRELESS EARPHONE THAT TRANSITIONS BETWEEN WIRELESS NETWORKS
US8407166B2 (en) * 2008-06-12 2013-03-26 Numenta, Inc. Hierarchical temporal memory system with higher-order temporal pooling capability
US8631351B2 (en) * 2008-06-29 2014-01-14 Microsoft Corporation Providing multiple degrees of context for content consumed on computers and media players
CA2730413A1 (en) * 2008-07-11 2010-01-14 Bby Solutions, Inc. Ratings switch for portable media players
US20100010997A1 (en) * 2008-07-11 2010-01-14 Abo Enterprise, LLC Method and system for rescoring a playlist
GB2457968A (en) * 2008-08-06 2009-09-02 John W Hannay & Co Ltd Forming a presentation of content
US7991757B2 (en) * 2008-08-12 2011-08-02 Amazon Technologies, Inc. System for obtaining recommendations from multiple recommenders
US7991650B2 (en) 2008-08-12 2011-08-02 Amazon Technologies, Inc. System for obtaining recommendations from multiple recommenders
US8694666B2 (en) * 2008-08-21 2014-04-08 Apple Inc. Personalized streaming digital content
JP5293018B2 (en) * 2008-09-09 2013-09-18 ソニー株式会社 Music information processing apparatus, music information processing method, and computer program
US20100076778A1 (en) * 2008-09-25 2010-03-25 Kondrk Robert H Method and System for Providing and Maintaining Limited-Subscriptions to Digital Media Assets
US8392505B2 (en) * 2008-09-26 2013-03-05 Apple Inc. Collaborative playlist management
US8707181B2 (en) * 2008-11-12 2014-04-22 Apple Inc. Preview of next media object to play
US8195582B2 (en) * 2009-01-16 2012-06-05 Numenta, Inc. Supervision based grouping of patterns in hierarchical temporal memory (HTM)
US9281847B2 (en) 2009-02-27 2016-03-08 Qualcomm Incorporated Mobile reception of digital video broadcasting—terrestrial services
US8473367B2 (en) * 2009-05-11 2013-06-25 Rahul U. Shah Computer based media access method and system
US20110029928A1 (en) * 2009-07-31 2011-02-03 Apple Inc. System and method for displaying interactive cluster-based media playlists
US8492638B2 (en) * 2009-08-05 2013-07-23 Robert Bosch Gmbh Personalized entertainment system
US20110035033A1 (en) * 2009-08-05 2011-02-10 Fox Mobile Dictribution, Llc. Real-time customization of audio streams
US9288010B2 (en) 2009-08-19 2016-03-15 Qualcomm Incorporated Universal file delivery methods for providing unequal error protection and bundled file delivery services
US20110096828A1 (en) * 2009-09-22 2011-04-28 Qualcomm Incorporated Enhanced block-request streaming using scalable encoding
US9917874B2 (en) 2009-09-22 2018-03-13 Qualcomm Incorporated Enhanced block-request streaming using block partitioning or request controls for improved client-side handling
US8612313B2 (en) * 2010-03-03 2013-12-17 Verizon Patent And Licensing Inc. Metadata subscription systems and methods
US20110225496A1 (en) * 2010-03-12 2011-09-15 Peter Jeffe Suggested playlist
EP2548096B1 (en) 2010-03-15 2017-10-25 Numenta, Inc. Temporal memory using sparse distributed representation
US11651277B2 (en) 2010-03-15 2023-05-16 Numenta, Inc. Sparse distributed representation for networked processing in predictive system
US9485546B2 (en) 2010-06-29 2016-11-01 Qualcomm Incorporated Signaling video samples for trick mode video representations
US8918533B2 (en) 2010-07-13 2014-12-23 Qualcomm Incorporated Video switching for streaming video data
US9185439B2 (en) 2010-07-15 2015-11-10 Qualcomm Incorporated Signaling data for multiplexing video components
KR101775027B1 (en) * 2010-07-21 2017-09-06 삼성전자주식회사 Method and apparatus for sharing content
US20120023403A1 (en) * 2010-07-21 2012-01-26 Tilman Herberger System and method for dynamic generation of individualized playlists according to user selection of musical features
US9596447B2 (en) 2010-07-21 2017-03-14 Qualcomm Incorporated Providing frame packing type information for video coding
US8806050B2 (en) 2010-08-10 2014-08-12 Qualcomm Incorporated Manifest file updates for network streaming of coded multimedia data
WO2012057744A1 (en) * 2010-10-27 2012-05-03 Hewlett-Packard Development Company, L.P. Providing control over a personalized category of information
JP5594532B2 (en) * 2010-11-09 2014-09-24 ソニー株式会社 Information processing apparatus and method, information processing system, and program
US9171318B2 (en) * 2010-11-15 2015-10-27 Verizon Patent And Licensing Inc. Virtual insertion of advertisements
US9158775B1 (en) * 2010-12-18 2015-10-13 Google Inc. Scoring stream items in real time
US11429343B2 (en) 2011-01-25 2022-08-30 Sonos, Inc. Stereo playback configuration and control
US11265652B2 (en) 2011-01-25 2022-03-01 Sonos, Inc. Playback device pairing
US8958375B2 (en) 2011-02-11 2015-02-17 Qualcomm Incorporated Framing for an improved radio link protocol including FEC
US9270299B2 (en) 2011-02-11 2016-02-23 Qualcomm Incorporated Encoding and decoding using elastic codes with flexible source block mapping
US9026942B2 (en) * 2011-02-25 2015-05-05 Cbs Interactive Inc. Song lyric processing with user interaction
FR2973134B1 (en) * 2011-03-23 2015-09-11 Xilopix METHOD FOR REFINING THE RESULTS OF A SEARCH IN A DATABASE
US20120254256A1 (en) * 2011-03-29 2012-10-04 Bose Corporation Multimedia Analysis and Cluster Formation
US8840013B2 (en) * 2011-12-06 2014-09-23 autoGraph, Inc. Consumer self-profiling GUI, analysis and rapid information presentation tools
US20140278992A1 (en) * 2013-03-15 2014-09-18 Nfluence Media, Inc. Ad blocking tools for interest-graph driven personalization
US8572553B2 (en) 2011-06-10 2013-10-29 International Business Machines Corporation Systems and methods for providing feedback for software components
US8504570B2 (en) 2011-08-25 2013-08-06 Numenta, Inc. Automated search for detecting patterns and sequences in data using a spatial and temporal memory system
US8645291B2 (en) 2011-08-25 2014-02-04 Numenta, Inc. Encoding of data for processing in a spatial and temporal memory system
US8825565B2 (en) 2011-08-25 2014-09-02 Numenta, Inc. Assessing performance in a spatial and temporal memory system
US8954448B1 (en) 2011-08-31 2015-02-10 Amazon Technologies, Inc. Presenting content related to current media consumption
US9253233B2 (en) 2011-08-31 2016-02-02 Qualcomm Incorporated Switch signaling methods providing improved switching between representations for adaptive HTTP streaming
CN102347945A (en) * 2011-09-22 2012-02-08 深圳市五巨科技有限公司 Method and system for feeding back abnormality of mobile terminal music player
CA3111501C (en) * 2011-09-26 2023-09-19 Sirius Xm Radio Inc. System and method for increasing transmission bandwidth efficiency ("ebt2")
US9843844B2 (en) 2011-10-05 2017-12-12 Qualcomm Incorporated Network streaming of media data
US8825668B2 (en) * 2011-11-16 2014-09-02 Google Inc. Method and apparatus for updating song playlists based on received user ratings
US9665339B2 (en) 2011-12-28 2017-05-30 Sonos, Inc. Methods and systems to select an audio track
US9292526B2 (en) 2012-02-24 2016-03-22 Empire Technology Development Llc Context-based content list generation
US9294226B2 (en) 2012-03-26 2016-03-22 Qualcomm Incorporated Universal object delivery and template-based file delivery
US8788659B1 (en) 2012-03-29 2014-07-22 Google Inc. Playlist analytics
US10225610B2 (en) * 2012-03-29 2019-03-05 Sony Interactive Entertainment LLC Method and apparatus for content channels using user feedback
US9659093B1 (en) 2012-04-02 2017-05-23 Google Inc. Adaptive recommendations of user-generated mediasets
US9729115B2 (en) 2012-04-27 2017-08-08 Sonos, Inc. Intelligently increasing the sound level of player
US20140214917A1 (en) 2012-05-23 2014-07-31 Clear Channel Management Services, Inc. Custom Voice Track
US9369514B2 (en) * 2012-06-08 2016-06-14 Spotify Ab Systems and methods of selecting content items
US10620797B2 (en) * 2012-06-29 2020-04-14 Spotify Ab Systems and methods for multi-context media control and playback
US8751429B2 (en) 2012-07-09 2014-06-10 Wine Ring, Inc. Personal taste assessment method and system
US8843951B1 (en) * 2012-08-27 2014-09-23 Google Inc. User behavior indicator
US9367587B2 (en) * 2012-09-07 2016-06-14 Pandora Media System and method for combining inputs to generate and modify playlists
US9008330B2 (en) 2012-09-28 2015-04-14 Sonos, Inc. Crossover frequency adjustments for audio speakers
US20140122593A1 (en) * 2012-10-16 2014-05-01 Apple Inc. Dynamically updating a shared radio station
US9244586B2 (en) * 2012-10-16 2016-01-26 Apple Inc. Displaying a buy/download button based on purchase history
US20140123005A1 (en) * 2012-10-25 2014-05-01 Apple Inc. User interface for streaming media stations with virtual playback
US9159021B2 (en) 2012-10-23 2015-10-13 Numenta, Inc. Performing multistep prediction using spatial and temporal memory system
US20140123004A1 (en) * 2012-10-25 2014-05-01 Apple Inc. Station creation
US10558729B2 (en) 2012-11-30 2020-02-11 Microsoft Technology Licensing, Llc Enhanced collection environments
US20140229894A1 (en) * 2013-02-12 2014-08-14 Daisy, Llc Systems and methods for generating playlists based on user feedback
AU2013378061A1 (en) 2013-02-14 2015-08-20 Ringit, Inc. Recommendation system based on group profiles of personal taste
US9154535B1 (en) * 2013-03-08 2015-10-06 Scott C. Harris Content delivery system with customizable content
US9335818B2 (en) 2013-03-15 2016-05-10 Pandora Media System and method of personalizing playlists using memory-based collaborative filtering
US10275463B2 (en) 2013-03-15 2019-04-30 Slacker, Inc. System and method for scoring and ranking digital content based on activity of network users
EP3036657A4 (en) * 2013-03-15 2017-06-21 Spotify AB Systems, methods, and computer readable medium for generating playlists
US9626438B2 (en) 2013-04-24 2017-04-18 Leaf Group Ltd. Systems and methods for determining content popularity based on searches
US11743544B2 (en) 2013-04-25 2023-08-29 Trent R McKenzie Interactive content feedback system
US11003708B2 (en) 2013-04-25 2021-05-11 Trent R. McKenzie Interactive music feedback system
US10102224B2 (en) * 2013-04-25 2018-10-16 Trent R. McKenzie Interactive music feedback system
US11423077B2 (en) 2013-04-25 2022-08-23 Trent R. McKenzie Interactive music feedback system
US9389754B2 (en) 2013-05-14 2016-07-12 Demand Media, Inc. Generating a playlist based on content meta data and user parameters
US10296639B2 (en) * 2013-09-05 2019-05-21 International Business Machines Corporation Personalized audio presentation of textual information
EP3074935A1 (en) * 2013-11-28 2016-10-05 Faulwetter, Patrick Platform device for providing quantitative collective knowledge
US20160378872A1 (en) 2013-11-28 2016-12-29 Patrick Faulwetter Platform Apparatus for Providing Qualitative Collective Knowledge
US9226087B2 (en) 2014-02-06 2015-12-29 Sonos, Inc. Audio output balancing during synchronized playback
US9226073B2 (en) 2014-02-06 2015-12-29 Sonos, Inc. Audio output balancing during synchronized playback
US9241198B2 (en) * 2014-02-14 2016-01-19 Surewaves Mediatech Private Limited Method and system for automatically scheduling and inserting television commercial and real-time updating of electronic program guide
US10032477B2 (en) * 2014-02-27 2018-07-24 Rovi Guides, Inc. Systems and methods for modifying a playlist of media assets based on user interactions with a playlist menu
US10380649B2 (en) 2014-03-03 2019-08-13 Spotify Ab System and method for logistic matrix factorization of implicit feedback data, and application to media environments
US10872110B2 (en) 2014-03-03 2020-12-22 Spotify Ab Systems, apparatuses, methods and computer-readable medium for automatically generating playlists based on taste profiles
US10448075B2 (en) 2014-03-06 2019-10-15 Cox Communications, Inc. Content conditioning and distribution of conditioned media assets at a content platform
US10318878B2 (en) 2014-03-19 2019-06-11 Numenta, Inc. Temporal processing scheme and sensorimotor information processing
WO2015145421A1 (en) * 2014-03-23 2015-10-01 Eran Bida Webcasting station platform
US10129599B2 (en) 2014-04-28 2018-11-13 Sonos, Inc. Media preference database
US9672213B2 (en) 2014-06-10 2017-06-06 Sonos, Inc. Providing media items from playback history
US10614724B2 (en) 2014-06-17 2020-04-07 Johnson & Johnson Consumer Inc. Systems and methods for wellness, health, and lifestyle planning, tracking, and maintenance
US10209947B2 (en) 2014-07-23 2019-02-19 Sonos, Inc. Device grouping
US9671997B2 (en) 2014-07-23 2017-06-06 Sonos, Inc. Zone grouping
US20160066038A1 (en) * 2014-08-28 2016-03-03 Dozo LLP Clip List Generation
US10346754B2 (en) * 2014-09-18 2019-07-09 Sounds Like Me Limited Method and system for psychological evaluation based on music preferences
US20160188290A1 (en) * 2014-12-30 2016-06-30 Anhui Huami Information Technology Co., Ltd. Method, device and system for pushing audio
US10516917B2 (en) * 2015-03-10 2019-12-24 Turner Broadcasting System, Inc. Providing a personalized entertainment network
US20160285937A1 (en) * 2015-03-24 2016-09-29 Spotify Ab Playback of streamed media content
US10298636B2 (en) * 2015-05-15 2019-05-21 Pandora Media, Llc Internet radio song dedication system and method
US10338794B2 (en) * 2015-05-15 2019-07-02 Spotify Ab Playback of media streams in dependence of a time of a day
US10248376B2 (en) 2015-06-11 2019-04-02 Sonos, Inc. Multiple groupings in a playback system
EP4270214A3 (en) 2015-09-01 2023-12-13 Dream It Get IT Limited Media unit retrieval and related processes
CN105227355B (en) * 2015-09-08 2018-07-06 广东欧珀移动通信有限公司 A kind of multimedia play list management method, apparatus and system
US9798823B2 (en) 2015-11-17 2017-10-24 Spotify Ab System, methods and computer products for determining affinity to a content creator
US10623518B2 (en) * 2016-02-04 2020-04-14 Spotify Ab System and method for ordering media content for shuffled playback based on user preference
US10860646B2 (en) 2016-08-18 2020-12-08 Spotify Ab Systems, methods, and computer-readable products for track selection
US9940390B1 (en) 2016-09-27 2018-04-10 Microsoft Technology Licensing, Llc Control system using scoped search and conversational interface
US10712997B2 (en) 2016-10-17 2020-07-14 Sonos, Inc. Room association based on name
US10888783B2 (en) 2017-09-20 2021-01-12 Sony Interactive Entertainment Inc. Dynamic modification of audio playback in games
US10661175B2 (en) 2017-09-26 2020-05-26 Sony Interactive Entertainment Inc. Intelligent user-based game soundtrack
CN107844587B (en) * 2017-11-16 2022-02-22 百度在线网络技术(北京)有限公司 Method and apparatus for updating multimedia playlist
US11097078B2 (en) 2018-09-26 2021-08-24 Cary Kochman Method and system for facilitating the transition between a conscious and unconscious state
US10936647B2 (en) 2018-10-04 2021-03-02 International Business Machines Corporation Generating and playing back media playlists via utilization of biometric and other data
EP3871118A4 (en) 2018-10-24 2023-03-22 Marketspringpad IP USA LLC. Methods and systems for interactive data management
US10893347B2 (en) * 2019-04-19 2021-01-12 Pio Networking Pte. Ltd. Intelligent audio playback and community sharing system
US11636855B2 (en) 2019-11-11 2023-04-25 Sonos, Inc. Media content based on operational data
US11681922B2 (en) 2019-11-26 2023-06-20 Numenta, Inc. Performing inference and training using sparse neural network
US11330315B2 (en) 2020-02-21 2022-05-10 Rovi Guides, Inc. Systems and methods for generating adapted content depictions
US11682002B2 (en) * 2020-06-05 2023-06-20 Marketspringpad Ip Usa Llc. Methods and systems for interactive data management
WO2023147184A1 (en) * 2022-01-31 2023-08-03 MuzicSwipe LLC Systems and methods for matching listeners and artists

Citations (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4843562A (en) * 1987-06-24 1989-06-27 Broadcast Data Systems Limited Partnership Broadcast information classification system and method
US5325298A (en) * 1990-11-07 1994-06-28 Hnc, Inc. Methods for generating or revising context vectors for a plurality of word stems
US5619709A (en) * 1993-09-20 1997-04-08 Hnc, Inc. System and method of context vector generation and retrieval
US5719344A (en) * 1995-04-18 1998-02-17 Texas Instruments Incorporated Method and system for karaoke scoring
US5728960A (en) * 1996-07-10 1998-03-17 Sitrick; David H. Multi-dimensional transformation systems and display communication architecture for musical compositions
US6018738A (en) * 1998-01-22 2000-01-25 Microsft Corporation Methods and apparatus for matching entities and for predicting an attribute of an entity based on an attribute frequency value
US20010002851A1 (en) * 1995-04-14 2001-06-07 Takao Shimada Multimedia data processing system in network
US20020002899A1 (en) * 2000-03-22 2002-01-10 Gjerdingen Robert O. System for content based music searching
US6353825B1 (en) * 1999-07-30 2002-03-05 Verizon Laboratories Inc. Method and device for classification using iterative information retrieval techniques
US20020032019A1 (en) * 2000-04-24 2002-03-14 Marks Michael B. Method for assembly of unique playlists
US20020069215A1 (en) * 2000-02-14 2002-06-06 Julian Orbanes Apparatus for viewing information in virtual space using multiple templates
US20020075311A1 (en) * 2000-02-14 2002-06-20 Julian Orbanes Method for viewing information in virtual space
US20020083060A1 (en) * 2000-07-31 2002-06-27 Wang Avery Li-Chun System and methods for recognizing sound and music signals in high noise and distortion
US6505154B1 (en) * 1999-02-13 2003-01-07 Primasoft Gmbh Method and device for comparing acoustic input signals fed into an input device with acoustic reference signals stored in a memory
US6539395B1 (en) * 2000-03-22 2003-03-25 Mood Logic, Inc. Method for creating a database for comparing music
US6542869B1 (en) * 2000-05-11 2003-04-01 Fuji Xerox Co., Ltd. Method for automatic analysis of audio including music and speech
US6633845B1 (en) * 2000-04-07 2003-10-14 Hewlett-Packard Development Company, L.P. Music summarization system and method
US6678679B1 (en) * 2000-10-10 2004-01-13 Science Applications International Corporation Method and system for facilitating the refinement of data queries
US6850252B1 (en) * 1999-10-05 2005-02-01 Steven M. Hoffberg Intelligent electronic appliance system and method
US6931451B1 (en) * 1996-10-03 2005-08-16 Gotuit Media Corp. Systems and methods for modifying broadcast programming
US6933433B1 (en) * 2000-11-08 2005-08-23 Viacom, Inc. Method for producing playlists for personalized music stations and for transmitting songs on such playlists
US20070163425A1 (en) * 2000-03-13 2007-07-19 Tsui Chi-Ying Melody retrieval system
US7272593B1 (en) * 1999-01-26 2007-09-18 International Business Machines Corporation Method and apparatus for similarity retrieval from iterative refinement
US7277766B1 (en) * 2000-10-24 2007-10-02 Moodlogic, Inc. Method and system for analyzing digital audio files
US8073854B2 (en) * 2007-04-10 2011-12-06 The Echo Nest Corporation Determining the similarity of music using cultural and acoustic information
US8326584B1 (en) * 1999-09-14 2012-12-04 Gracenote, Inc. Music searching methods based on human perception
US8429026B1 (en) * 1999-06-28 2013-04-23 Dietfood Corp. System and method for creating and submitting electronic shopping lists
US8996540B2 (en) * 2005-12-19 2015-03-31 Apple Inc. User to user recommender
US20230130778A1 (en) * 2021-01-28 2023-04-27 Adobe Inc. Generating visual data stories
US20230134240A1 (en) * 2016-10-31 2023-05-04 Rovi Guides, Inc. Systems and methods for flexibly using trending topics as parameters for recommending media assets that are related to a viewed media asset
US20230161811A1 (en) * 2020-07-07 2023-05-25 Huawei Technologies Co., Ltd. Image search system, method, and apparatus

Family Cites Families (330)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3947882A (en) 1971-06-01 1976-03-30 Lightner Robert W Vending system for remotely accessible stored information
US3935379A (en) 1974-05-09 1976-01-27 General Dynamics Corporation Method of and system for adaptive run length encoding of image representing digital information
US4061890A (en) 1976-08-27 1977-12-06 Bose Corporation Loudspeaker with single layer rectangular wire voice coil wound on slit metal bobbin with a notch in the adjacent pole plate
US4577069A (en) 1976-08-27 1986-03-18 Bose Corporation Electroacoustical transducer
US4158756A (en) 1976-08-27 1979-06-19 Bose Corporation Dynamic loudspeaker with plastic basket encapsulating front pole plate
USRE31679E (en) * 1976-09-02 1984-09-18 Bose Corporation Loudspeaker enclosure with acoustical compliance resonating with port mass at frequency below system resonance
US4227220A (en) 1977-05-02 1980-10-07 Hill Brown Tape recording system
US4191472A (en) 1977-10-17 1980-03-04 Derek Mason Apparatus for the elevation of coins
NL8103064A (en) 1981-06-25 1983-01-17 Philips Nv COMMON AERIAL DEVICE FOR THE RECEPTION AND DISTRIBUTION OF TV AND DIGITAL AUDIO SIGNALS.
US4788675A (en) 1983-10-03 1988-11-29 Jones Markley L Music delivery system
US4549631A (en) 1983-10-24 1985-10-29 Bose Corporation Multiple porting loudspeaker systems
US4546342A (en) 1983-12-14 1985-10-08 Digital Recording Research Limited Partnership Data compression method and apparatus
US4891633A (en) 1984-07-23 1990-01-02 General Research Of Electronics, Inc. Digital image exchange system
US4729016A (en) 1985-05-06 1988-03-01 Eastman Kodak Company Digital color image processing method and apparatus employing three color reproduction functions for adjusting both tone scale and color balance
US4792968A (en) 1985-07-10 1988-12-20 Fdr Interactive Technologies Statistical analysis system for use with public communication facility
US4761642A (en) 1985-10-04 1988-08-02 Tektronix, Inc. System for providing data communication between a computer terminal and a plurality of concurrent processes running on a multiple process computer
US4703465A (en) 1985-12-04 1987-10-27 1K Entertainment Center Ltd. Method and apparatus for producing an audio magnetic tape recording from a preselected music library
US4789980A (en) 1986-08-18 1988-12-06 American Telephone & Telegraph Company, At&T Bell Laboratories Switching techniques for FDM communication systems
US4775935A (en) 1986-09-22 1988-10-04 Westinghouse Electric Corp. Video merchandising system with variable and adoptive product sequence presentation order
US4870515A (en) 1986-12-05 1989-09-26 Stokes Richard A Music memory data recording, storage and playback system for magnetic recording and/or reproducing apparatus
US4739514A (en) 1986-12-22 1988-04-19 Bose Corporation Automatic dynamic equalizing
US4852910A (en) 1987-01-05 1989-08-01 Jon Taffer Apparatus and method for selecting and playing music
US5365381A (en) 1987-01-06 1994-11-15 Magic Music Cassette Company High speed recorded information duplicating equipment
JP2658030B2 (en) 1987-01-30 1997-09-30 ソニー株式会社 Information transmission equipment
US4851931A (en) 1987-02-20 1989-07-25 1K Music International Ltd. Method and apparatus for producing an audio magnetic tape recording at high speed from a preselected music library
US4939714A (en) 1987-02-27 1990-07-03 Pioneer Electronic Corporation Method for fading in and out reproduced digitally recorded data
US5394521A (en) 1991-12-09 1995-02-28 Xerox Corporation User interface with multiple workspaces for sharing display system objects
US5072412A (en) 1987-03-25 1991-12-10 Xerox Corporation User interface with multiple workspaces for sharing display system objects
US4780758A (en) 1987-04-30 1988-10-25 Gte Government Systems Corporation Telecommunication system with burst and continuous audio signals
US4996642A (en) 1987-10-01 1991-02-26 Neonics, Inc. System and method for recommending items
US4937807A (en) 1987-10-15 1990-06-26 Personics Corporation System for encoding sound recordings for high-density storage and high-speed transfers
US4974173A (en) 1987-12-02 1990-11-27 Xerox Corporation Small-scale workspace representations indicating activities by other users
US4885747A (en) 1988-02-17 1989-12-05 International Business Machines Corp. Broadband and baseband LAN
US5041972A (en) 1988-04-15 1991-08-20 Frost W Alan Method of measuring and evaluating consumer response for the development of consumer products
US5124911A (en) 1988-04-15 1992-06-23 Image Engineering, Inc. Method of evaluating consumer choice through concept testing for the marketing and development of consumer products
US4890320A (en) 1988-06-09 1989-12-26 Monslow H Vincent Television broadcast system for selective transmission of viewer-chosen programs at viewer-requested times
US5191573A (en) 1988-06-13 1993-03-02 Hair Arthur R Method for transmitting a desired digital video or audio signal
US5121478A (en) 1988-09-08 1992-06-09 Xerox Corporation Window system with independently replaceable window functionality
US4975690A (en) 1988-11-07 1990-12-04 Ibm Corporation Method for concurrent data entry and manipulation in multiple applications
US5001554A (en) 1988-12-23 1991-03-19 Scientific-Atlanta, Inc. Terminal authorization method
US5272700A (en) 1989-02-28 1993-12-21 First Pacific Networks, Inc. Spectrally efficient broadband transmission system
US5274836A (en) 1989-08-08 1993-12-28 Gde Systems, Inc. Multiple encoded carrier data link
US5168481A (en) 1989-12-29 1992-12-01 Culbertson Charles R Automated digital broadcast system
DE4190031B4 (en) 1990-01-18 2005-04-14 E-MU Systems, Inc., Scotts Valley Data compression of decaying musical instrument sounds for a digital scanning system
US5210820A (en) 1990-05-02 1993-05-11 Broadcast Data Systems Limited Partnership Signal recognition system and method
US5355302A (en) 1990-06-15 1994-10-11 Arachnid, Inc. System for managing a plurality of computer jukeboxes
US5781889A (en) 1990-06-15 1998-07-14 Martin; John R. Computer jukebox and jukebox network
KR920009642B1 (en) 1990-08-24 1992-10-22 삼성전자 주식회사 Data condensing circuit
US5237157A (en) 1990-09-13 1993-08-17 Intouch Group, Inc. Kiosk apparatus and method for point of preview and for compilation of market data
US5963916A (en) 1990-09-13 1999-10-05 Intouch Group, Inc. Network apparatus and method for preview of music products and compilation of market data
US5054360A (en) 1990-11-01 1991-10-08 International Business Machines Corporation Method and apparatus for simultaneous output of digital audio and midi synthesized music
JP2605960B2 (en) 1990-11-29 1997-04-30 松下電器産業株式会社 Video transmission equipment
US5092424A (en) 1990-12-03 1992-03-03 Bose Corporation Electroacoustical transducing with at least three cascaded subchambers
US5351234A (en) 1990-12-28 1994-09-27 Nynex Corporation System for integrated distribution of switched voice and television on coaxial cable
US5132992A (en) 1991-01-07 1992-07-21 Paul Yurt Audio and video transmission and receiving system
US5131048A (en) 1991-01-09 1992-07-14 Square D Company Audio distribution system
JPH04241381A (en) 1991-01-16 1992-08-28 Brother Ind Ltd Electronic learning machine
US5291395A (en) 1991-02-07 1994-03-01 Max Abecassis Wallcoverings storage and retrieval system
US5634101A (en) 1991-05-31 1997-05-27 R. Alan Blau & Associates, Co. Method and apparatus for obtaining consumer information
JPH0535287A (en) 1991-07-31 1993-02-12 Ricos:Kk 'karaoke' music selection device
US5634021A (en) 1991-08-15 1997-05-27 Borland International, Inc. System and methods for generation of design images based on user design inputs
US5278751A (en) 1991-08-30 1994-01-11 International Business Machines Corporation Dynamic manufacturing process control
DE4132004A1 (en) * 1991-09-26 1993-04-01 Broadcast Television Syst METHOD FOR INDUCTIVE TRANSMISSION OF A DATA SIGNAL
JPH05100699A (en) 1991-10-11 1993-04-23 Sharp Corp Sound recording and reproducing device
US20010013123A1 (en) 1991-11-25 2001-08-09 Freeman Michael J. Customized program creation by splicing server based video, audio, or graphical segments
US5445295A (en) 1992-01-17 1995-08-29 Brown; Graham Automated vending machine system for recorded goods
US5267865A (en) 1992-02-11 1993-12-07 John R. Lee Interactive computer aided natural learning method and apparatus
US5446891A (en) 1992-02-26 1995-08-29 International Business Machines Corporation System for adjusting hypertext links with weighed user goals and activities
US5469206A (en) 1992-05-27 1995-11-21 Philips Electronics North America Corporation System and method for automatically correlating user preferences with electronic shopping information
US5594601A (en) 1992-05-27 1997-01-14 Tandy Corporation Method of programming audio tracks in a sequential medium
US5303302A (en) 1992-06-18 1994-04-12 Digital Equipment Corporation Network packet receiver with buffer logic for reassembling interleaved data packets
US5371551A (en) 1992-10-29 1994-12-06 Logan; James Time delayed digital video system using concurrent recording and playback
US5508984A (en) 1992-10-30 1996-04-16 Kabushiki Kaisha Kenwood Optical disc player having program title display
US5542071A (en) 1992-11-13 1996-07-30 Video Associates Labs, Inc. System for determining communication speed of parallel printer port of computer by using start timer and stop timer commands within data combined with embedded strobe
US5333135A (en) 1993-02-01 1994-07-26 North American Philips Corporation Identification of a data stream transmitted as a sequence of packets
US5719786A (en) 1993-02-03 1998-02-17 Novell, Inc. Digital media data stream network management system
US5444489A (en) 1993-02-11 1995-08-22 Georgia Tech Research Corporation Vector quantization video encoder using hierarchical cache memory scheme
US5565910A (en) 1993-03-26 1996-10-15 Vionx, Inc. Data and television network for digital computer workstations
US6186794B1 (en) 1993-04-02 2001-02-13 Breakthrough To Literacy, Inc. Apparatus for interactive adaptive learning by an individual through at least one of a stimuli presentation device and a user perceivable display
US5679911A (en) 1993-05-26 1997-10-21 Pioneer Electronic Corporation Karaoke reproducing apparatus which utilizes data stored on a recording medium to make the apparatus more user friendly
US5485221A (en) 1993-06-07 1996-01-16 Scientific-Atlanta, Inc. Subscription television system and terminal for enabling simultaneous display of multiple services
KR0183682B1 (en) 1993-06-30 1999-04-15 김광호 A karaoke tv
KR0165271B1 (en) 1993-06-30 1999-03-20 김광호 Method for deciding medley function to a television
US5483276A (en) * 1993-08-02 1996-01-09 The Arbitron Company Compliance incentives for audience monitoring/recording devices
US5418713A (en) 1993-08-05 1995-05-23 Allen; Richard Apparatus and method for an on demand data delivery system for the preview, selection, retrieval and reproduction at a remote location of previously recorded or programmed materials
US5761601A (en) * 1993-08-09 1998-06-02 Nemirofsky; Frank R. Video distribution of advertisements to businesses
US5408630A (en) 1993-08-20 1995-04-18 Iii G.I. Ltd. Three-state virtual volume system for managing document storage to permanent media
DE4331258A1 (en) 1993-09-15 1995-03-23 Nsm Ag Program-controlled entertainment and play equipment
US5594726A (en) 1993-09-17 1997-01-14 Scientific-Atlanta, Inc. Frequency agile broadband communications system
US5469219A (en) 1993-09-20 1995-11-21 Micrognosis, Inc. RF-based high-bandwidth signal transmission method and apparatus
US5410344A (en) 1993-09-22 1995-04-25 Arrowsmith Technologies, Inc. Apparatus and method of selecting video programs based on viewers' preferences
US5485518A (en) 1993-09-30 1996-01-16 Yellowstone Environmental Science, Inc. Electronic media program recognition and choice
JPH07175868A (en) 1993-10-15 1995-07-14 Internatl Business Mach Corp <Ibm> Method and system for output of digital information to medium
CA2175187A1 (en) 1993-10-28 1995-05-04 William K. Thomson Database search summary with user determined characteristics
JPH07202730A (en) 1993-12-28 1995-08-04 Matsushita Electric Ind Co Ltd Signal transmitter
US5408259A (en) 1993-12-30 1995-04-18 Northern Telecom Limited Data modulation arrangement for selectively distributing data
US5629867A (en) 1994-01-25 1997-05-13 Goldman; Robert J. Selection and retrieval of music from a digital database
US5594792A (en) 1994-01-28 1997-01-14 American Telecorp Methods and apparatus for modeling and emulating devices in a network of telecommunication systems
AU1889895A (en) * 1994-03-08 1995-09-25 Preben Alstrom A neural network
US5619425A (en) 1994-03-22 1997-04-08 Brother Kogyo Kabushiki Kaisha Data transmission system
US5608446A (en) 1994-03-31 1997-03-04 Lucent Technologies Inc. Apparatus and method for combining high bandwidth and low bandwidth data transfer
DE4413451A1 (en) 1994-04-18 1995-12-14 Rolf Brugger Device for the distribution of music information in digital form
US5724567A (en) 1994-04-25 1998-03-03 Apple Computer, Inc. System for directing relevance-ranked data objects to computer users
US5550575A (en) 1994-05-04 1996-08-27 West; Brett Viewer discretion television program control system
US5592511A (en) 1994-05-10 1997-01-07 Schoen; Neil C. Digital customized audio products with user created data and associated distribution and production system
US6769009B1 (en) 1994-05-31 2004-07-27 Richard R. Reisman Method and system for selecting a personalized set of information channels
US5610893A (en) 1994-06-02 1997-03-11 Olympus Optical Co., Ltd. Information recording and reproducing apparatus for copying information from exchangeable master recording medium to a plurality of other exchangeable recording media
KR100395040B1 (en) 1994-06-09 2003-12-31 코닌클리케 필립스 일렉트로닉스 엔.브이. Interactive multi-access communication system, and the central and user stations used in this system
US5754773A (en) 1994-06-16 1998-05-19 Lucent Technologies, Inc. Multimedia on-demand server having different transfer rates
US5541638A (en) 1994-06-28 1996-07-30 At&T Corp. User programmable entertainment method and apparatus
US6134547A (en) 1994-07-11 2000-10-17 Muze, Inc. Computerized method and system for user-interactive, multimedia cataloguing, navigation and previewing of film and films on video
US5557541A (en) 1994-07-21 1996-09-17 Information Highway Media Corporation Apparatus for distributing subscription and on-demand audio programming
US5583868A (en) 1994-07-25 1996-12-10 Microsoft Corporation Method and system for combining data from multiple servers into a single continuous data stream using a switch
US5623652A (en) 1994-07-25 1997-04-22 Apple Computer, Inc. Method and apparatus for searching for information in a network and for controlling the display of searchable information on display devices in the network
US5564002A (en) 1994-08-01 1996-10-08 International Business Machines Corporation Method and apparatus for implementing a virtual desktop through window positioning
US5566353A (en) * 1994-09-06 1996-10-15 Bylon Company Limited Point of purchase video distribution system
US5668948A (en) 1994-09-08 1997-09-16 International Business Machines Corporation Media streamer with control node enabling same isochronous streams to appear simultaneously at output ports or different streams to appear simultaneously at output ports
US5573953A (en) 1994-09-09 1996-11-12 Quantum Group, Inc. Method for enhancing the response of a biomimetic sensor
US5481509A (en) 1994-09-19 1996-01-02 Software Control Systems, Inc. Jukebox entertainment system including removable hard drives
CA2158588C (en) 1994-09-21 2003-01-14 Belwinder Singh Barn Multiple channel terminal server communications network
US5652627A (en) 1994-09-27 1997-07-29 Lucent Technologies Inc. System and method for reducing jitter in a packet-based transmission network
US5781734A (en) 1994-09-28 1998-07-14 Nec Corporation System for providing audio and video services on demand
EP0786121B1 (en) 1994-10-12 2000-01-12 Touchtunes Music Corporation Intelligent digital audiovisual playback system
JPH10507853A (en) 1994-10-25 1998-07-28 オブジェクト テクノロジー ライセンシング コーポレイション Object oriented system for servicing windows
US5611607A (en) 1994-10-28 1997-03-18 Elf Technologies Corporation Electronic librarian and filing system and method
US5534911A (en) 1994-11-02 1996-07-09 Levitan; Gutman Virtual personal channel in a television system
US5696965A (en) 1994-11-03 1997-12-09 Intel Corporation Electronic information appraisal agent
US5717923A (en) 1994-11-03 1998-02-10 Intel Corporation Method and apparatus for dynamically customizing electronic information to individual end users
WO1996017306A2 (en) 1994-11-21 1996-06-06 Oracle Corporation Media server
US5870723A (en) 1994-11-28 1999-02-09 Pare, Jr.; David Ferrin Tokenless biometric transaction authorization method and system
US6029195A (en) 1994-11-29 2000-02-22 Herz; Frederick S. M. System for customized electronic identification of desirable objects
US5623613A (en) 1994-11-29 1997-04-22 Microsoft Corporation System for displaying programming information
US5758257A (en) 1994-11-29 1998-05-26 Herz; Frederick System and method for scheduling broadcast of and access to video programs and other data using customer profiles
US5793980A (en) 1994-11-30 1998-08-11 Realnetworks, Inc. Audio-on-demand communication system
US5550746A (en) 1994-12-05 1996-08-27 American Greetings Corporation Method and apparatus for storing and selectively retrieving product data by correlating customer selection criteria with optimum product designs based on embedded expert judgments
EP0716370A3 (en) 1994-12-06 2005-02-16 International Business Machines Corporation A disk access method for delivering multimedia and video information on demand over wide area networks
US5541919A (en) 1994-12-19 1996-07-30 Motorola, Inc. Multimedia multiplexing device and method using dynamic packet segmentation
US5768280A (en) 1995-01-12 1998-06-16 Way; Winston I. Two-way broadband CATV communication network protocol and monitoring technology
US5652749A (en) 1995-02-03 1997-07-29 International Business Machines Corporation Apparatus and method for segmentation and time synchronization of the transmission of a multiple program multimedia data stream
CA2683230C (en) 1995-02-13 2013-08-27 Intertrust Technologies Corporation Systems and methods for secure transaction management and electronic rights protection
US5619250A (en) 1995-02-19 1997-04-08 Microware Systems Corporation Operating system for interactive television system set top box utilizing dynamic system upgrades
US6288991B1 (en) 1995-03-06 2001-09-11 Fujitsu Limited Storage medium playback method and device
US5855015A (en) 1995-03-20 1998-12-29 Interval Research Corporation System and method for retrieval of hyperlinked information resources
US5896179A (en) 1995-03-31 1999-04-20 Cirrus Logic, Inc. System for displaying computer generated images on a television set
US5635980A (en) 1995-04-04 1997-06-03 Bell Communications Research, Inc. System and method for customer premises broadband interface with on-hook alerting
US5749081A (en) 1995-04-06 1998-05-05 Firefly Network, Inc. System and method for recommending items to a user
US5616876A (en) 1995-04-19 1997-04-01 Microsoft Corporation System and methods for selecting music on the basis of subjective content
US5586121A (en) 1995-04-21 1996-12-17 Hybrid Networks, Inc. Asymmetric hybrid access system and method
US5708961A (en) 1995-05-01 1998-01-13 Bell Atlantic Network Services, Inc. Wireless on-premises video distribution using digital multiplexing
US5861906A (en) 1995-05-05 1999-01-19 Microsoft Corporation Interactive entertainment network system and method for customizing operation thereof according to viewer preferences
US5674734A (en) * 1995-05-18 1997-10-07 President And Fellows Of Harvard College Cell death protein
US6070160A (en) 1995-05-19 2000-05-30 Artnet Worldwide Corporation Non-linear database set searching apparatus and method
US5625608A (en) 1995-05-22 1997-04-29 Lucent Technologies Inc. Remote control device capable of downloading content information from an audio system
US5670730A (en) 1995-05-22 1997-09-23 Lucent Technologies Inc. Data protocol and method for segmenting memory for a music chip
US5841979A (en) 1995-05-25 1998-11-24 Information Highway Media Corp. Enhanced delivery of audio data
US5675784A (en) 1995-05-31 1997-10-07 International Business Machnes Corporation Data structure for a relational database system for collecting component and specification level data related to products
US5701451A (en) 1995-06-07 1997-12-23 International Business Machines Corporation Method for fulfilling requests of a web browser
US6181867B1 (en) 1995-06-07 2001-01-30 Intervu, Inc. Video storage and retrieval system
US5721878A (en) 1995-06-07 1998-02-24 International Business Machines Corporation Multimedia control system and method for controlling multimedia program presentation
US5682550A (en) 1995-06-07 1997-10-28 International Business Machines Corporation System for restricting user access to default work area settings upon restoration from nonvolatile memory where the settings are independent of the restored state information
US5740549A (en) 1995-06-12 1998-04-14 Pointcast, Inc. Information and advertising distribution system and method
US6041311A (en) 1995-06-30 2000-03-21 Microsoft Corporation Method and apparatus for item recommendation using automated collaborative filtering
US5825829A (en) 1995-06-30 1998-10-20 Scientific-Atlanta, Inc. Modulator for a broadband communications system
US5659539A (en) 1995-07-14 1997-08-19 Oracle Corporation Method and apparatus for frame accurate access of digital audio-visual information
US5822524A (en) 1995-07-21 1998-10-13 Infovalue Computing, Inc. System for just-in-time retrieval of multimedia files over computer networks by transmitting data packets at transmission rate determined by frame size
US5751672A (en) 1995-07-26 1998-05-12 Sony Corporation Compact disc changer utilizing disc database
US5781449A (en) 1995-08-10 1998-07-14 Advanced System Technologies, Inc. Response time measurement apparatus and method
US6026388A (en) 1995-08-16 2000-02-15 Textwise, Llc User interface and other enhancements for natural language information retrieval system and method
EP0935251B1 (en) 1995-08-21 2007-01-03 Matsushita Electric Industrial Co., Ltd. Method for reproducing a multimedia optical disc which enables a title developer to coordinate the use of special reproduction functions
EP0762766A3 (en) 1995-09-12 1997-11-05 AT&T Corp. Network apparatus and method for providing two-way broadband communications
GB2305339B (en) 1995-09-14 2000-03-29 Ian Mark Ackroyd Apparatus and a method for accessing data
US5911131A (en) 1995-12-20 1999-06-08 Vig; Tommy Computer aided calculation, appraisal and valuation of works of art
US5648628A (en) 1995-09-29 1997-07-15 Ng; Tao Fei S. Cartridge supported karaoke device
US5708845A (en) 1995-09-29 1998-01-13 Wistendahl; Douglass A. System for mapping hot spots in media content for interactive digital media program
US5798923A (en) 1995-10-18 1998-08-25 Intergraph Corporation Optimal projection design and analysis
US5737747A (en) 1995-10-27 1998-04-07 Emc Corporation Prefetching to service multiple video streams from an integrated cached disk array
US5722041A (en) 1995-12-05 1998-02-24 Altec Lansing Technologies, Inc. Hybrid home-entertainment system
US5726909A (en) 1995-12-08 1998-03-10 Krikorian; Thomas M. Continuous play background music system
US5745095A (en) 1995-12-13 1998-04-28 Microsoft Corporation Compositing digital information on a display screen based on screen descriptor
US5745685A (en) 1995-12-29 1998-04-28 Mci Communications Corporation Protocol extension in NSPP using an acknowledgment bit
US6091725A (en) 1995-12-29 2000-07-18 Cisco Systems, Inc. Method for traffic management, traffic prioritization, access control, and packet forwarding in a datagram computer network
JP3112411B2 (en) 1996-01-11 2000-11-27 松下電器産業株式会社 Video transmission equipment
US6088005A (en) 1996-01-11 2000-07-11 Hewlett-Packard Company Design and method for a large, virtual workspace
US6009465A (en) 1996-01-22 1999-12-28 Svi Systems, Inc. Entertainment and information systems and related management networks for a remote video delivery system
US5695400A (en) 1996-01-30 1997-12-09 Boxer Jam Productions Method of managing multi-player game playing over a network
US5760822A (en) 1996-01-30 1998-06-02 Lucent Technologies Inc. Central node converter for local network having single coaxial cable
US5754771A (en) 1996-02-12 1998-05-19 Sybase, Inc. Maximum receive capacity specifying query processing client/server system replying up to the capacity and sending the remainder upon subsequent request
GB9602809D0 (en) 1996-02-12 1996-04-10 Northern Telecom Ltd A bidirectional communications network
US5864868A (en) 1996-02-13 1999-01-26 Contois; David C. Computer control system and user interface for media playing devices
US6047327A (en) 1996-02-16 2000-04-04 Intel Corporation System for distributing electronic information to a targeted group of users
US5633839A (en) 1996-02-16 1997-05-27 Alexander; Gregory Music vending machine capable of recording a customer's music selections onto a compact disc
US6018768A (en) 1996-03-08 2000-01-25 Actv, Inc. Enhanced video programming system and method for incorporating and displaying retrieved integrated internet information segments
US5893095A (en) 1996-03-29 1999-04-06 Virage, Inc. Similarity engine for content-based retrieval of images
US5850340A (en) 1996-04-05 1998-12-15 York; Matthew Integrated remote controlled computer and television system
US6134590A (en) 1996-04-16 2000-10-17 Webtv Networks, Inc. Method and apparatus for automatically connecting devices to a local network
US5960445A (en) 1996-04-24 1999-09-28 Sony Corporation Information processor, method of updating a program and information processing system
US5848396A (en) 1996-04-26 1998-12-08 Freedom Of Information, Inc. Method and apparatus for determining behavioral profile of a computer user
US5850433A (en) 1996-05-01 1998-12-15 Sprint Communication Co. L.P. System and method for providing an on-line directory service
US5945988A (en) 1996-06-06 1999-08-31 Intel Corporation Method and apparatus for automatically determining and dynamically updating user preferences in an entertainment system
US5912676A (en) 1996-06-14 1999-06-15 Lsi Logic Corporation MPEG decoder frame memory interface which is reconfigurable for different frame store architectures
US6064379A (en) 1996-06-24 2000-05-16 Sun Microsystems, Inc. System and method for synchronizing presentation of media stream playlists with real time
US5929850A (en) 1996-07-01 1999-07-27 Thomson Consumer Electronices, Inc. Interactive television system and method having on-demand web-like navigational capabilities for displaying requested hyperlinked web-like still images associated with television content
JPH1021261A (en) 1996-07-05 1998-01-23 Hitachi Ltd Method and system for multimedia data base retrieval
US5822543A (en) 1996-07-08 1998-10-13 International Business Machines Corporation Gathering data handling statistics in non-synchronous data communication networks
US5822744A (en) 1996-07-15 1998-10-13 Kesel; Brad Consumer comment reporting apparatus and method
US6038610A (en) 1996-07-17 2000-03-14 Microsoft Corporation Storage of sitemaps at server sites for holding information regarding content
US5918223A (en) 1996-07-22 1999-06-29 Muscle Fish Method and article of manufacture for content-based analysis, storage, retrieval, and segmentation of audio information
US5913204A (en) 1996-08-06 1999-06-15 Kelly; Thomas L. Method and apparatus for surveying music listener opinion about songs
US5991374A (en) 1996-08-08 1999-11-23 Hazenfield; Joey C. Programmable messaging system for controlling playback of messages on remote music on-hold- compatible telephone systems and other message output devices
US5740134A (en) 1996-08-13 1998-04-14 Peterson; Tim Musical CD creation unit
US6353822B1 (en) 1996-08-22 2002-03-05 Massachusetts Institute Of Technology Program-listing appendix
US5745938A (en) * 1996-08-30 1998-05-05 Westvaco Corporation Rescue board
US5890152A (en) 1996-09-09 1999-03-30 Seymour Alvin Rapaport Personal feedback browser for obtaining media files
US5926624A (en) 1996-09-12 1999-07-20 Audible, Inc. Digital information library and delivery system with logic for generating files targeted to the playback device
US5819160A (en) 1996-09-18 1998-10-06 At&T Corp Programmable radio subscription system for receiving selectively defined information
FR2753868A1 (en) 1996-09-25 1998-03-27 Technical Maintenance Corp METHOD FOR SELECTING A RECORDING ON AN AUDIOVISUAL DIGITAL REPRODUCTION SYSTEM AND SYSTEM FOR IMPLEMENTING THE METHOD
US6018343A (en) 1996-09-27 2000-01-25 Timecruiser Computing Corp. Web calendar architecture and uses thereof
JP2943910B2 (en) * 1996-09-30 1999-08-30 日本電気株式会社 Drive device for piezoelectric transformer
US5732216A (en) 1996-10-02 1998-03-24 Internet Angles, Inc. Audio message exchange system
US6199076B1 (en) 1996-10-02 2001-03-06 James Logan Audio program player including a dynamic program selection controller
US5897639A (en) 1996-10-07 1999-04-27 Greef; Arthur Reginald Electronic catalog system and method with enhanced feature-based search
US5983200A (en) 1996-10-09 1999-11-09 Slotznick; Benjamin Intelligent agent for executing delegated tasks
US5889949A (en) 1996-10-11 1999-03-30 C-Cube Microsystems Processing system with memory arbitrating between memory access requests in a set top box
US5923385A (en) 1996-10-11 1999-07-13 C-Cube Microsystems Inc. Processing system with single-buffered display capture
US5781227A (en) 1996-10-25 1998-07-14 Diva Systems Corporation Method and apparatus for masking the effects of latency in an interactive information distribution system
US5801747A (en) 1996-11-15 1998-09-01 Hyundai Electronics America Method and apparatus for creating a television viewer profile
US6031795A (en) 1996-12-02 2000-02-29 Thomson Consumer Electronics, Inc. Method and apparatus for programming a jukebox with information related to content on media contained therein
US5835727A (en) 1996-12-09 1998-11-10 Sun Microsystems, Inc. Method and apparatus for controlling access to services within a computer network
US5931901A (en) 1996-12-09 1999-08-03 Robert L. Wolfe Programmed music on demand from the internet
EP0945017A1 (en) 1996-12-12 1999-09-29 Conexant Systems, Inc. Digital video converter box for subscriber/home with multiple television sets
US5734119A (en) 1996-12-19 1998-03-31 Invision Interactive, Inc. Method for streaming transmission of compressed music
US5739451A (en) 1996-12-27 1998-04-14 Franklin Electronic Publishers, Incorporated Hand held electronic music encyclopedia with text and note structure search
US6324267B1 (en) 1997-01-17 2001-11-27 Scientific-Atlanta, Inc. Two-tiered authorization and authentication for a cable data delivery system
CA2279631A1 (en) 1997-01-29 1998-07-30 West-Walker, Francis Nigel Method of transferring media files over a communications network
US6014706A (en) 1997-01-30 2000-01-11 Microsoft Corporation Methods and apparatus for implementing control functions in a streamed video display system
US5907843A (en) 1997-02-27 1999-05-25 Apple Computer, Inc. Replaceable and extensible navigator component of a network component system
US6031818A (en) 1997-03-19 2000-02-29 Lucent Technologies Inc. Error correction system for packet switching networks
US5848404A (en) 1997-03-24 1998-12-08 International Business Machines Corporation Fast query search in large dimension database
US6182122B1 (en) 1997-03-26 2001-01-30 International Business Machines Corporation Precaching data at an intermediate server based on historical data requests by users of the intermediate server
US5926207A (en) 1997-03-31 1999-07-20 Compaq Computer Corporation Channel server functionality
US5959945A (en) 1997-04-04 1999-09-28 Advanced Technology Research Sa Cv System for selectively distributing music to a plurality of jukeboxes
US5987525A (en) 1997-04-15 1999-11-16 Cddb, Inc. Network delivery of interactive entertainment synchronized to playback of audio recordings
US5963957A (en) 1997-04-28 1999-10-05 Philips Electronics North America Corporation Bibliographic music data base with normalized musical themes
US6243725B1 (en) 1997-05-21 2001-06-05 Premier International, Ltd. List building system
US5920856A (en) 1997-06-09 1999-07-06 Xerox Corporation System for selecting multimedia databases over networks
US6017219A (en) 1997-06-18 2000-01-25 International Business Machines Corporation System and method for interactive reading and language instruction
US5983218A (en) 1997-06-30 1999-11-09 Xerox Corporation Multimedia database for use over networks
US6230200B1 (en) 1997-09-08 2001-05-08 Emc Corporation Dynamic modeling for resource allocation in a file server
US5960411A (en) 1997-09-12 1999-09-28 Amazon.Com, Inc. Method and system for placing a purchase order via a communications network
US6125387A (en) 1997-09-30 2000-09-26 The United States Of America Represented By The Secretary Of The Navy Operating methods for robust computer systems permitting autonomously switching between alternative/redundant
US6026398A (en) 1997-10-16 2000-02-15 Imarket, Incorporated System and methods for searching and matching databases
US5982363A (en) 1997-10-24 1999-11-09 General Instrument Corporation Personal computer-based set-top converter for television services
US6060997A (en) 1997-10-27 2000-05-09 Motorola, Inc. Selective call device and method for providing a stream of information
US6029165A (en) 1997-11-12 2000-02-22 Arthur Andersen Llp Search and retrieval information system and method
US5973683A (en) 1997-11-24 1999-10-26 International Business Machines Corporation Dynamic regulation of television viewing content based on viewer profile and viewing history
ES2397501T3 (en) * 1997-11-25 2013-03-07 Motorola Mobility, Llc Methods, systems and manufactured elements of audio content reproduction
US6009422A (en) 1997-11-26 1999-12-28 International Business Machines Corporation System and method for query translation/semantic translation using generalized query language
US6741869B1 (en) 1997-12-12 2004-05-25 International Business Machines Corporation Radio-like appliance for receiving information from the internet
US5986200A (en) 1997-12-15 1999-11-16 Lucent Technologies Inc. Solid state interactive music playback device
US6588015B1 (en) 1998-01-14 2003-07-01 General Instrument Corporation Broadcast interactive digital radio
US20020002039A1 (en) 1998-06-12 2002-01-03 Safi Qureshey Network-enabled audio device
US6385596B1 (en) 1998-02-06 2002-05-07 Liquid Audio, Inc. Secure online music distribution system
US6108686A (en) 1998-03-02 2000-08-22 Williams, Jr.; Henry R. Agent-based on-line information retrieval and viewing system
US7185355B1 (en) 1998-03-04 2007-02-27 United Video Properties, Inc. Program guide system with preference profiles
US6269275B1 (en) 1998-03-31 2001-07-31 Michael G. Slade Method and system for customizing and distributing presentations for user sites
US6118450A (en) 1998-04-03 2000-09-12 Sony Corporation Graphic user interface that is usable as a PC interface and an A/V interface
US6049797A (en) 1998-04-07 2000-04-11 Lucent Technologies, Inc. Method, apparatus and programmed medium for clustering databases with categorical attributes
US6240423B1 (en) 1998-04-22 2001-05-29 Nec Usa Inc. Method and system for image querying using region based and boundary based image matching
US6446080B1 (en) 1998-05-08 2002-09-03 Sony Corporation Method for creating, modifying, and playing a custom playlist, saved as a virtual CD, to be played by a digital audio/visual actuator device
US6093880A (en) 1998-05-26 2000-07-25 Oz Interactive, Inc. System for prioritizing audio for a virtual environment
US6031797A (en) 1998-06-11 2000-02-29 Sony Corporation Media playback device capable of shuffled playback based on a user's preferences
US7231175B2 (en) 1998-06-16 2007-06-12 United Video Properties, Inc. Music information system for obtaining information on a second music program while a first music program is played
US6953886B1 (en) 1998-06-17 2005-10-11 Looney Productions, Llc Media organizer and entertainment center
US5969283A (en) 1998-06-17 1999-10-19 Looney Productions, Llc Music organizer and entertainment center
FR2781582B1 (en) 1998-07-21 2001-01-12 Technical Maintenance Corp SYSTEM FOR DOWNLOADING OBJECTS OR FILES FOR SOFTWARE UPDATE
US6983371B1 (en) * 1998-10-22 2006-01-03 International Business Machines Corporation Super-distribution of protected digital content
US6611812B2 (en) * 1998-08-13 2003-08-26 International Business Machines Corporation Secure electronic content distribution on CDS and DVDs
US6389403B1 (en) * 1998-08-13 2002-05-14 International Business Machines Corporation Method and apparatus for uniquely identifying a customer purchase in an electronic distribution system
US6226618B1 (en) * 1998-08-13 2001-05-01 International Business Machines Corporation Electronic content delivery system
AU1704900A (en) 1998-10-13 2000-05-01 Radiowave.Com, Inc. System and method for determining the audience of digital radio programmes broadcast through the internet
US6223210B1 (en) 1998-10-14 2001-04-24 Radio Computing Services, Inc. System and method for an automated broadcast system
US6313833B1 (en) 1998-10-16 2001-11-06 Prophet Financial Systems Graphical data collection and retrieval interface
US6820063B1 (en) 1998-10-26 2004-11-16 Microsoft Corporation Controlling access to content based on certificates and access predicates
US6314094B1 (en) 1998-10-29 2001-11-06 Central Coast Patent Agency Inc Mobile wireless internet portable radio
KR100321437B1 (en) 1998-11-02 2002-05-13 이상훈 Mp3 player for vehicles
US6804825B1 (en) 1998-11-30 2004-10-12 Microsoft Corporation Video on demand methods and systems
US6813711B1 (en) 1999-01-05 2004-11-02 Samsung Electronics Co., Ltd. Downloading files from approved web site
US20050210101A1 (en) 1999-03-04 2005-09-22 Universal Electronics Inc. System and method for providing content, management, and interactivity for client devices
US6338044B1 (en) 1999-03-17 2002-01-08 Loudeye Technologies, Inc. Personal digital content system
US6434621B1 (en) 1999-03-31 2002-08-13 Hannaway & Associates Apparatus and method of using the same for internet and intranet broadcast channel creation and management
US6411724B1 (en) 1999-07-02 2002-06-25 Koninklijke Philips Electronics N.V. Using meta-descriptors to represent multimedia information
US7072932B1 (en) 1999-08-26 2006-07-04 Lucent Technologies Inc. Personalized network-based services
US6774926B1 (en) 1999-09-03 2004-08-10 United Video Properties, Inc. Personal television channel system
US6351736B1 (en) 1999-09-03 2002-02-26 Tomer Weisberg System and method for displaying advertisements with played data
US7181691B2 (en) 1999-09-16 2007-02-20 Sharp Laboratories Of America, Inc. Audiovisual information management system with presentation service
US7022905B1 (en) 1999-10-18 2006-04-04 Microsoft Corporation Classification of information and use of classifications in searching and retrieval of information
US7072846B1 (en) 1999-11-16 2006-07-04 Emergent Music Llc Clusters for rapid artist-audience matching
US6192340B1 (en) 1999-10-19 2001-02-20 Max Abecassis Integration of music from a personal library with real-time information
US6526411B1 (en) 1999-11-15 2003-02-25 Sean Ward System and method for creating dynamic playlists
US7213005B2 (en) * 1999-12-09 2007-05-01 International Business Machines Corporation Digital content distribution using web broadcasting services
WO2001047273A1 (en) 1999-12-21 2001-06-28 Tivo, Inc. Intelligent system and methods of recommending media content items based on user preferences
US20020010759A1 (en) * 1999-12-30 2002-01-24 Hitson Bruce L. System and method for multimedia content composition and distribution
US6434747B1 (en) 2000-01-19 2002-08-13 Individual Network, Inc. Method and system for providing a customized media list
US6662195B1 (en) 2000-01-21 2003-12-09 Microstrategy, Inc. System and method for information warehousing supporting the automatic, real-time delivery of personalized informational and transactional data to users via content delivery device
US6993290B1 (en) 2000-02-11 2006-01-31 International Business Machines Corporation Portable personal radio system and method
US6248946B1 (en) * 2000-03-01 2001-06-19 Ijockey, Inc. Multimedia content delivery system and method
US7275256B1 (en) 2001-08-28 2007-09-25 Music Choice System and method for providing an interactive, visual complement to an audio program
US7320025B1 (en) 2002-03-18 2008-01-15 Music Choice Systems and methods for providing a broadcast entertainment service and an on-demand entertainment service
US7028082B1 (en) 2001-03-08 2006-04-11 Music Choice Personalized audio system and method
US7325043B1 (en) 2000-03-08 2008-01-29 Music Choice System and method for providing a personalized media service
US7856485B2 (en) 2000-03-08 2010-12-21 Music Choice Systems and methods for providing customized media channels
US7133924B1 (en) 2000-03-08 2006-11-07 Music Choice Personalized audio system and method
US6879963B1 (en) 2000-04-12 2005-04-12 Music Choice Cross channel delivery system and method
US8271341B2 (en) 2000-04-12 2012-09-18 Music Choice Media content delivery systems and methods
JP2001306581A (en) * 2000-04-18 2001-11-02 Sony Corp Middleware and media data audiovisual equipment using the middleware
US8352331B2 (en) * 2000-05-03 2013-01-08 Yahoo! Inc. Relationship discovery engine
US7075000B2 (en) 2000-06-29 2006-07-11 Musicgenome.Com Inc. System and method for prediction of musical preferences
AU2001267815A1 (en) 2000-06-29 2002-01-08 Musicgenome.Com Inc. Using a system for prediction of musical preferences for the distribution of musical content over cellular networks
US6657117B2 (en) 2000-07-14 2003-12-02 Microsoft Corporation System and methods for providing automatic classification of media entities according to tempo properties
JP2002114107A (en) 2000-10-10 2002-04-16 Nissan Motor Co Ltd Audio equipment and method for playing music
US6819963B2 (en) 2000-12-06 2004-11-16 Advanced Micro Devices, Inc. Run-to-run control method for proportional-integral-derivative (PID) controller tuning for rapid thermal processing (RTP)
US6937770B1 (en) * 2000-12-28 2005-08-30 Emc Corporation Adaptive bit rate control for rate reduction of MPEG coded video
US6771568B2 (en) * 2001-01-12 2004-08-03 Sima Products Corporation Digital audio recorder
EP1241588A3 (en) * 2001-01-23 2006-01-04 Matsushita Electric Industrial Co., Ltd. Audio information provision system
US6993532B1 (en) 2001-05-30 2006-01-31 Microsoft Corporation Auto playlist generator
US20030236582A1 (en) 2002-06-25 2003-12-25 Lee Zamir Selection of items based on user reactions
US7158169B1 (en) 2003-03-07 2007-01-02 Music Choice Method and system for displaying content while reducing burn-in of a display
JP4713129B2 (en) 2004-11-16 2011-06-29 ソニー株式会社 Music content playback device, music content playback method, and music content and attribute information recording device
US7689556B2 (en) 2005-01-31 2010-03-30 France Telecom Content navigation service
US7756388B2 (en) 2005-03-21 2010-07-13 Microsoft Corporation Media item subgroup generation from a library

Patent Citations (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4843562A (en) * 1987-06-24 1989-06-27 Broadcast Data Systems Limited Partnership Broadcast information classification system and method
US5325298A (en) * 1990-11-07 1994-06-28 Hnc, Inc. Methods for generating or revising context vectors for a plurality of word stems
US5619709A (en) * 1993-09-20 1997-04-08 Hnc, Inc. System and method of context vector generation and retrieval
US5794178A (en) * 1993-09-20 1998-08-11 Hnc Software, Inc. Visualization of information using graphical representations of context vector based relationships and attributes
US20010002851A1 (en) * 1995-04-14 2001-06-07 Takao Shimada Multimedia data processing system in network
US5719344A (en) * 1995-04-18 1998-02-17 Texas Instruments Incorporated Method and system for karaoke scoring
US5728960A (en) * 1996-07-10 1998-03-17 Sitrick; David H. Multi-dimensional transformation systems and display communication architecture for musical compositions
US6931451B1 (en) * 1996-10-03 2005-08-16 Gotuit Media Corp. Systems and methods for modifying broadcast programming
US6018738A (en) * 1998-01-22 2000-01-25 Microsft Corporation Methods and apparatus for matching entities and for predicting an attribute of an entity based on an attribute frequency value
US7272593B1 (en) * 1999-01-26 2007-09-18 International Business Machines Corporation Method and apparatus for similarity retrieval from iterative refinement
US6505154B1 (en) * 1999-02-13 2003-01-07 Primasoft Gmbh Method and device for comparing acoustic input signals fed into an input device with acoustic reference signals stored in a memory
US8429026B1 (en) * 1999-06-28 2013-04-23 Dietfood Corp. System and method for creating and submitting electronic shopping lists
US6353825B1 (en) * 1999-07-30 2002-03-05 Verizon Laboratories Inc. Method and device for classification using iterative information retrieval techniques
US8326584B1 (en) * 1999-09-14 2012-12-04 Gracenote, Inc. Music searching methods based on human perception
US6850252B1 (en) * 1999-10-05 2005-02-01 Steven M. Hoffberg Intelligent electronic appliance system and method
US20020069215A1 (en) * 2000-02-14 2002-06-06 Julian Orbanes Apparatus for viewing information in virtual space using multiple templates
US20020075311A1 (en) * 2000-02-14 2002-06-20 Julian Orbanes Method for viewing information in virtual space
US20070163425A1 (en) * 2000-03-13 2007-07-19 Tsui Chi-Ying Melody retrieval system
US6539395B1 (en) * 2000-03-22 2003-03-25 Mood Logic, Inc. Method for creating a database for comparing music
US20020002899A1 (en) * 2000-03-22 2002-01-10 Gjerdingen Robert O. System for content based music searching
US6633845B1 (en) * 2000-04-07 2003-10-14 Hewlett-Packard Development Company, L.P. Music summarization system and method
US20020032019A1 (en) * 2000-04-24 2002-03-14 Marks Michael B. Method for assembly of unique playlists
US6542869B1 (en) * 2000-05-11 2003-04-01 Fuji Xerox Co., Ltd. Method for automatic analysis of audio including music and speech
US20020083060A1 (en) * 2000-07-31 2002-06-27 Wang Avery Li-Chun System and methods for recognizing sound and music signals in high noise and distortion
US6678679B1 (en) * 2000-10-10 2004-01-13 Science Applications International Corporation Method and system for facilitating the refinement of data queries
US7277766B1 (en) * 2000-10-24 2007-10-02 Moodlogic, Inc. Method and system for analyzing digital audio files
US6933433B1 (en) * 2000-11-08 2005-08-23 Viacom, Inc. Method for producing playlists for personalized music stations and for transmitting songs on such playlists
US8996540B2 (en) * 2005-12-19 2015-03-31 Apple Inc. User to user recommender
US8073854B2 (en) * 2007-04-10 2011-12-06 The Echo Nest Corporation Determining the similarity of music using cultural and acoustic information
US20230134240A1 (en) * 2016-10-31 2023-05-04 Rovi Guides, Inc. Systems and methods for flexibly using trending topics as parameters for recommending media assets that are related to a viewed media asset
US20230161811A1 (en) * 2020-07-07 2023-05-25 Huawei Technologies Co., Ltd. Image search system, method, and apparatus
US20230130778A1 (en) * 2021-01-28 2023-04-27 Adobe Inc. Generating visual data stories

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8903843B2 (en) * 2006-06-21 2014-12-02 Napo Enterprises, Llc Historical media recommendation service
US20150074090A1 (en) * 2006-06-21 2015-03-12 Napo Enterprises, Llc Historical Media Recommendation Service
US20090077052A1 (en) * 2006-06-21 2009-03-19 Concert Technology Corporation Historical media recommendation service
US11032290B2 (en) 2010-09-15 2021-06-08 Core Mobile Networks, Inc. Context-based analytics and intelligence
US20120066393A1 (en) * 2010-09-15 2012-03-15 Core Mobile Networks, Inc. System and method for real time delivery of context based content from the cloud to mobile devices
US8606923B2 (en) * 2010-09-15 2013-12-10 Core Mobile Networks, Inc. System and method for real time delivery of context based content from the cloud to mobile devices
US9071649B2 (en) 2010-09-15 2015-06-30 Core Mobile Networks, Inc. System and method for real time delivery of context based content from the cloud to mobile
KR20210146426A (en) * 2013-12-19 2021-12-03 그레이스노트, 인코포레이티드 Media service
US20220156308A1 (en) * 2013-12-19 2022-05-19 Gracenote, Inc. Station library creaton for a media service
KR20160101979A (en) * 2013-12-19 2016-08-26 그레이스노트, 인코포레이티드 Media service
US11868392B2 (en) * 2013-12-19 2024-01-09 Gracenote, Inc. Station library creation for a media service
US20160196270A1 (en) * 2013-12-19 2016-07-07 Gracenote, Inc. Station library creaton for a media service
KR102582911B1 (en) * 2013-12-19 2023-09-27 그레이스노트, 인코포레이티드 Media service
US10108619B2 (en) * 2013-12-19 2018-10-23 Gracenote, Inc. Station library creaton for a media service
KR20220104061A (en) * 2013-12-19 2022-07-25 그레이스노트, 인코포레이티드 Media service
KR102419851B1 (en) * 2013-12-19 2022-07-13 그레이스노트, 인코포레이티드 Media service
KR102331083B1 (en) * 2013-12-19 2021-11-25 그레이스노트, 인코포레이티드 Media service
WO2015094558A1 (en) * 2013-12-19 2015-06-25 Gracenote, Inc. Media service
US11269946B2 (en) 2013-12-19 2022-03-08 Gracenote, Inc. Station library creation for a media service
US9471572B1 (en) 2013-12-31 2016-10-18 Google Inc. Recommending candidates for consumption
US9619470B2 (en) 2014-02-04 2017-04-11 Google Inc. Adaptive music and video recommendations
WO2015120068A1 (en) * 2014-02-04 2015-08-13 Google Inc. Adaptive music and video recommendations
US10430452B2 (en) 2014-02-04 2019-10-01 Google Llc Adaptive music and video recommendations
CN105960640A (en) * 2014-02-04 2016-09-21 谷歌公司 Adaptive music and video recommendations
US20170124187A1 (en) * 2014-04-17 2017-05-04 Renault S.A.S. Method of adaptation of a representation of musical content

Also Published As

Publication number Publication date
US7962482B2 (en) 2011-06-14
WO2007067250A1 (en) 2007-06-14
US20110213769A1 (en) 2011-09-01
US20060212444A1 (en) 2006-09-21
US8306976B2 (en) 2012-11-06

Similar Documents

Publication Publication Date Title
US8306976B2 (en) Methods and systems for utilizing contextual feedback to generate and modify playlists
US20060212442A1 (en) Methods of Presenting and Providing Content to a User
US7797272B2 (en) System and method for dynamic playlist of media
US20060206478A1 (en) Playlist generating methods
US9164994B2 (en) Intelligent default weighting process for criteria utilized to score media content items
US9081780B2 (en) System and method for assigning user preference settings for a category, and in particular a media category
US8600838B2 (en) Website, user interfaces, and applications facilitating improved media search capability
US8175989B1 (en) Music recommendation system using a personalized choice set
US7370031B2 (en) Generation of high affinity media
JP5432264B2 (en) Apparatus and method for collection profile generation and communication based on collection profile
US8839141B2 (en) Method and system for visually indicating a replay status of media items on a media device
US20150269256A1 (en) System and method for cross-library recommendation
US9529801B2 (en) Multimedia analysis and recommendation generation
US20090055396A1 (en) Scoring and replaying media items
US20090144273A1 (en) System and method for music and compatibility matching
US20090049045A1 (en) Method and system for sorting media items in a playlist on a media device
WO2002037334A1 (en) System and method for performing content experience management
US20120254256A1 (en) Multimedia Analysis and Cluster Formation

Legal Events

Date Code Title Description
AS Assignment

Owner name: WELLS FARGO BANK, NATIONAL ASSOCIATION, CALIFORNIA

Free format text: SECURITY INTEREST;ASSIGNORS:PANDORA MEDIA, INC.;PANDORA MEDIA CALIFORNIA, LLC;REEL/FRAME:044985/0009

Effective date: 20171229

AS Assignment

Owner name: PANDORA MEDIA, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HANDMAN, ETIENNE F.;CONRAD, THOMAS J.;KENNEDY, JOSEPH J.;AND OTHERS;SIGNING DATES FROM 20060426 TO 20060527;REEL/FRAME:044756/0063

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

AS Assignment

Owner name: ADSWIZ INC., CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WELLS FARGO BANK, NATIONAL ASSOCIATION;REEL/FRAME:048209/0925

Effective date: 20190201

Owner name: PANDORA MEDIA CALIFORNIA, LLC, CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WELLS FARGO BANK, NATIONAL ASSOCIATION;REEL/FRAME:048209/0925

Effective date: 20190201

AS Assignment

Owner name: PANDORA MEDIA, LLC, CALIFORNIA

Free format text: CHANGE OF NAME;ASSIGNOR:PANDORA MEDIA, INC.;REEL/FRAME:048806/0776

Effective date: 20190201

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCV Information on status: appeal procedure

Free format text: NOTICE OF APPEAL FILED

STCV Information on status: appeal procedure

Free format text: APPEAL BRIEF (OR SUPPLEMENTAL BRIEF) ENTERED AND FORWARDED TO EXAMINER

STCV Information on status: appeal procedure

Free format text: EXAMINER'S ANSWER TO APPEAL BRIEF MAILED

STCV Information on status: appeal procedure

Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS

STCV Information on status: appeal procedure

Free format text: BOARD OF APPEALS DECISION RENDERED

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS