Music Streaming, Playlist Creation And Streaming Architecture

Patent No. US11899713 (titled "Music Streaming, Playlist Creation And Streaming Architecture") was filed by Muvox Llc on Jan 5, 2023.

What is this patent about?

’713 is related to the field of music streaming services, specifically addressing the challenge of music publishers providing streamed music to end users in a non-programmatic way, allowing for personalized playlists. Existing services often lack a way for publishers to directly engage users and ensure they are aware of the source of the music, while also struggling to efficiently categorize and recommend music based on objective criteria.

The underlying idea behind ’713 is to use computer-derived rhythm, texture, and pitch (RTP) scores to categorize music tracks. These scores are derived from analyzing low-level data extracted from the tracks, which are then mapped to high-level acoustic attributes. This allows for objective categorization and playlist creation based on musical characteristics rather than subjective human tagging. The system also aims to create a universal database of RTP scores to avoid redundant analysis by different music publishers.

The claims of ’713 focus on selecting a song based on a computer-derived comparison between a representation of the song and known similarities in representations of other songs. This comparison relies on a human-trained machine that uses representations of other songs, where the representation of the song is based on isolating and identifying frequency characteristics. The selection is based on the similarity between the moods of the song and the moods of the other songs.

In practice, the system involves extracting low-level data from music tracks, analyzing this data to develop high-level acoustic attributes, and then using these attributes to generate RTP scores. These scores are then used to categorize the tracks and create collections that can be accessed by end-user applications. A key aspect is the use of a universal database where RTP scores are stored, allowing different music publishers to share this information and avoid redundant analysis. This database is accessed through an API server, ensuring controlled access and data integrity.

The system differentiates itself from prior approaches by using objective, computer-derived RTP scores for music categorization, rather than relying solely on human tagging or genre classifications. The human-trained machine aspect is also important, as it allows the system to learn and improve its categorization accuracy over time. Furthermore, the universal database and API server architecture enables efficient sharing of RTP scores among different music publishers, reducing redundancy and improving the overall efficiency of the music streaming ecosystem.

How does this patent fit in bigger picture?

Technical landscape at the time

In the mid-2010s when ’713 was filed, music streaming services were prevalent, at a time when users typically interacted with these services through dedicated applications or web interfaces. Systems commonly relied on large catalogs of licensed music tracks, and hardware or software constraints made personalized playlist generation based on detailed audio analysis non-trivial.

Novelty and Inventive Step

The examiner allowed the application because the claims overcome prior art rejections. While prior art teaches selecting songs based on musical characteristics and filtering similar songs based on criteria like mood and frequency, it does not explicitly teach that similarities in digitized representations of songs are based on a human-trained machine listening to each song to isolate and identify frequency characteristics. The claims explicitly recite this feature.

Claims

This patent contains 20 claims, with independent claims 1 and 16. The independent claims are directed to a method and a system for selecting a song based on computer-derived comparisons of song representations and known similarities, considering frequency characteristics and moods. The dependent claims generally elaborate on the method and system, providing details about representations, audio fingerprints, frequency data analysis, mood correlation, and song attributes.

Key Claim Terms New

Definitions of key terms used in the patent claims.

Term (Source)Support for SpecificationInterpretation
Computer-derived comparison
(Claim 1)
“The playlists may be based on computer-derived rhythm, texture and pitch (RTP) scores for predetermined tracks. The RTP scores may be derived from high-level acoustic attributes for tracks determined by an analyzed extraction of low-level data from the tracks. The RTP scores may be used to determine a category for each track among the plurality of categories and playlists of the tracks may be based on the categories and other factors.”A comparison performed by a computer between a representation of a song and known similarities in representations of other songs.
Digitized representation
(Claim 16)
“For example, a chromagram image may be scanned from left to right, one pixel at a time, in grayscale, to create a large number of subimages. Then, characteristic filters may be applied to the subimages to capture intensity differences over time in the musical notes represented by the subimages. Each filter may calculate the sum of specific areas of the subimages, compare the sums, and arrange the areas in a number of ways.”A digital representation of sampled frequency data, filtered to capture intensity differences, corresponding to one or more attributes of a song.
Frequency characteristics
(Claim 1, Claim 16)
“With respect to mood, when someone listens to a track, the track tends to evoke some emotion. Since those inherent qualities may be represented by frequency-related data (i.e., frequencies, structure and organization), that frequency-related data may be used to identify those inherent qualities.”The frequency-related data of a song, used to determine its mood.
Human-trained machine
(Claim 1, Claim 16)
“A solution for analyzing spectrograms extracted from audio files may involve converting each audio file into mel-spectrograms, representing multiple audio features extracted from each predetermined period (such as one second) of each audio file, that are then input to a trained neural network. This neural network may be trained based on an ontology of audio event classes and a large-scale collection of human-labeled sound clips.”A machine learning model trained using data derived from human listening and analysis of songs.

Patent Family

Patent Family

File Wrapper

The dossier documents provide a comprehensive record of the patent's prosecution history - including filings, correspondence, and decisions made by patent offices - and are crucial for understanding the patent's legal journey and any challenges it may have faced during examination.

  • Date

    Description

  • Get instant alerts for new documents

US11899713

MUVOX LLC
Application Number
US18150728
Filing Date
Jan 5, 2023
Status
Granted
Expiry Date
Jan 22, 2035
External Links
Slate, USPTO, Google Patents