WO2015170126A1 - Methods, systems and computer program products for identifying commonalities of rhythm between disparate musical tracks and using that information to make music recommendations - Google Patents
Methods, systems and computer program products for identifying commonalities of rhythm between disparate musical tracks and using that information to make music recommendations Download PDFInfo
- Publication number
- WO2015170126A1 WO2015170126A1 PCT/GB2015/051380 GB2015051380W WO2015170126A1 WO 2015170126 A1 WO2015170126 A1 WO 2015170126A1 GB 2015051380 W GB2015051380 W GB 2015051380W WO 2015170126 A1 WO2015170126 A1 WO 2015170126A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- media content
- digital media
- metric
- track
- metadata
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/68—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/683—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/43—Querying
- G06F16/435—Filtering based on additional data, e.g. user or group profiles
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/68—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
Definitions
- the field of the invention relates to methods, systems and computer program products for identifying commonalities of rhythm between disparate musical tracks and using that information to make music recommendations.
- a common problem in the field of music provision is the problem of which music tracks to recommend, both in terms of making explicit recommendations to individual consumers based on their existing musical taste and in terms of selecting a musical track to be played after any given musical track has concluded.
- the present invention provides a method for analysing audio media to extract metadata describing the rhythm, tempo and other descriptive metadata about the audio track and utilising that metadata to assist in cross -matching tracks, providing recommendations, perform additional Digital Signal Processing of tracks and for sundry other purposes as disclosed in this document.
- the present invention teaches a method for efficiently and reliably identifying commonalities of rhythm between disparate musical tracks and using that information to make music recommendations accordingly.
- a computer program product embodied on a non-transitory storage medium or on a cellular mobile telephone device or on another hardware device, the computer program product operable to automatically generate recommendations for digital media content, in which that system is adapted to automatically generate recommendations of digital media content for an end user by (a) analysing a first item or set of digital media content and its associated metadata to calculate metric-based metadata for that digital media content and (b) using that analysis to identify other digital media content which have a specifically-defined relationship with that first digital media content; and (c) provide, to the end user, recommendations of one or more items of that digital media content which were identified in the previous step.
- the metric-based metadata comprises one or more of: the rhythmic value or values of a track, the time signature or time signatures of a track; the rhythmic simplicity or complexity of a track; the tempo or cadence drift of a track; the rhythm salience and rhythmicity; confidence values for any of the listed derived metrics; any other calculated metrics for a track.
- the method of point 12 where the subset of digital media content items is determined by filtering the list using additional criteria, including but not limited to the artist, album, genre, musical taste, current date, time of playback, date/time of most recent playback, user demographic data or by using any other additional criteria.
- a computer based system adapted to perform the method of any preceding point.
- Figure 1 shows a Tree representation for metrical hierarchy, (a) A simple duple hierarchy
- duple hierarchy dividing the lower level into three groups of two (e.g. 4 time signature).
- Figure 2 shows an example of a special case with two metrical levels that can't coexist in the same hierarchy, hence requiring the creation of two parallel hierarchies: usage of triplets and eight notes at the same time (a.k.a 3:2 polyrh thm) represented by f J+2 and f, res ectively with 1 the quarter note level. Then whereas
- Figure 3 shows elementary patterns and the sequence of amplitudes associated to both their plain and accented interpretations.
- Figure 4 shows Spectrum ⁇ if) w j m extracted metrical hierarchy H , metrical levels rates F anc
- music and “media content” in this document are to be taken to encompass all “media content” which is in digital form or which it is possible to convert to digital form - including but not limited to books, magazines, newspapers and other periodicals, video in the form of digital video, motion pictures, television shows (as series, as seasons and as individual episodes), computer games and other interactive media, images (photographic or otherwise) and music.
- track indicates a specific item of media content, whether that be a song, a television show, an eBook or portion thereof, a computer game or any other discreet item of media content.
- playlist “timeline” and “album” are used interchangeably to indicate collections of “tracks” and/or interstitials which have been conjoined together such that they may be treated as a single entity for the purposes of analysis or recommendation.
- digital media catalogue digital music catalogue
- media catalogue media catalogue
- catalog catalogs
- the terms “digital media catalogue”, “digital music catalogue”, “media catalogue” and “catalogue” are used interchangeably to indicate a collection of tracks and/ or albums to which a user may be allowed access for listening purposes.
- the digital media catalogue may aggregate both digital media files and their associated metadata or, in another example embodiment, the digital media and metadata may be delivered from multiple such catalogues. There is no implication that only one such catalogue exists, and the term encompasses access to multiple separate catalogues simultaneously, whether consecutively, concurrently or by aggregation.
- the actual catalogue utilised by any given operation may be fixed or may vary over time and/ or according to the location or access rights of a particular device or end-user.
- DRM Digital Rights Management
- the verbs "to listen”, “to view” and “to play” are to be taken as encompassing any interaction between a human and media content, whether that be listening to audio content, watching video or image content, reading books or other textual content, playing a computer game, interacting with interactive media content or some combination of such activities.
- the playing of digital media content is to be further construed as occurring where some or all of one digital media content file is incorporated within another digital media content file such that playing the latter would necessitate the former also being played.
- the terms “user”, “consumer”, “end user” and “individual” are used interchangeably to refer to the person, or group of people making use of the facilities provided by the interface. In all cases, the masculine includes the feminine and vice versa.
- device and “media player” are used interchangeably to refer to any computational device which is capable of playing digital media content, including but not limited to MP3 players, television sets, home entertainment system, home computer systems, mobile computing devices, games consoles, handheld games consoles, IVEs or other vehicular-based media players or any other applicable device or software media player on such a device. Something essentially capable of playback of media.
- DSP Digital Signal Processing
- Such calculated metadata may take a variety of forms, including deriving the tempo of a musical track or identifying one or more spots within the digital media file which are gauged to be representative of that content as a whole.
- hook is used to refer to one or more portions of a digital media file which have been identified, whether via DSP or manually or by some other method, as being representative of the content as a whole.
- a movie trailer consists of a series of one or more "hooks” from the movie while particularly apposite riffs or lines from a musical track serve a similar identifying purpose.
- partial playback and “extract” are used interchangeably to refer to the playback of an item of digital media content where that item is not played back in its entirety or is incorporated into another item of digital media content.
- BPM is used to denote the count of the number of Beats Per Minute in an audio track.
- FFT Fast Fourier Transform
- MIR Music Information Retrieval
- DJ disk Jockey
- DJ disk Jockey
- Tempo estimation has been an important aspect of Music Information Retrieval (MIR) research since its inception. A large body of research has been devoted to the task. Tempo is commonly considered as a fundamental property of the music, which certainly explains why it has been one of the first MIR tasks to be investigated.
- MIR Music Information Retrieval
- Tempo is often associated to our perception of the speed or pace of the music or, in other words, how fast we feel the music is. Its relevance for music databases navigation, recommendation systems or automatic curation is then obvious.
- tempo can be represented by a single rate (frequency).
- tempo is commonly associated with speed or pace of music, that is to say our sense of how fast the music is is governed by one rate.
- a nuance between metronomic indication and tempo need to be made here as are generally represented in the form of a BPM value.
- a metronomic indication is written on a music score and associates a note value with an absolute frequency measurement, thereby providing an absolute point of reference for speed of execution - usually using a metronome. This has several implications. First, a metronomic indication makes sense only in regard to the score it is attached to.
- the "correct" metronomic indication for a given score is not necessarily unique. Numerous examples of this can be found in music history, Chopin being an example of a composer who used different notations on different editions of his scores.
- a metronomic indication is relevant for performers, not necessarily for listeners. The score is a cue to the performers so that they can deliver the musical feel intended by the composer (including the sense of pace) but it does not mean a single element of the score (the metronomic indication) bears this information.
- Italian tempo indications such as presto, allegro, lento provide information which would be a lot more meaningful to listeners than just a metronomic indication since they describe what the music should sound like rather than the prescription of a technical performance element.
- ground truth tempo is obtained from the score and effectively refers to the metronomic indication.
- the access to, or even the existence of, a music score with a metronomic indication is not easy, and is furthermore subject to copyright law. For that reason, unless working on symbolic data, this source of ground truth tends not to be commonly used.
- the "perceived tempo” ground truth data is typically collected by asking people to tap along a music piece, record the hits and measure the tapping rate. This is what the ground truth data used for the MIREX audio tempo estimation task consists of.
- the present invention provides a descriptor that is more closely related to the temporal feel of music than a single rate (whether it is called tempo or tactus).
- a descriptor has direct applications in the context of today's music distribution and consumption such as content-based recommendation for streaming and download services, automatic radio, curation assistance etc.
- the metrical structure appears to play an important role in perception of speed: differences in surface rhythm and metrical structure do interfere with judgments of tempo, even if two passages have the same beat rate. Consequently, an important focus of the present invention is on metrical structure extraction from audio. It is also safe to assume that the metrical structure influences more that just the sense of speed. For instance, music in a compound meter feels different from a simple meter.
- Rhythm patterns in music are complex signals. In music which can be considered to have a sense of meter, the rhythm will have multiple related periodicities forming a hierarchical structure; the common musical terms for which include bars, beats and beat subdivisions.
- the present invention introduces a mid-level feature which captures the weighting of salient pulse rates present in the music, which correspond to the most likely metrical hierarchy.
- Feature extraction is achieved by computing a novelty curve from the audio then processing this in time and frequency to produce a rhythmic spectrogram or rhythmogram. From this, we derive multiple possible candidates for a metrical hierarchy that fits the rhythmic content the best. That feature provides a meaningful description of the rhythm content without applying specific labels, such as tempo or tactus, to any particular level.
- the temporal occurrences of musical events (e.g. notes) in a musical piece can be, and often are, organised such that periodic patterns emerge. These can be decomposed into different levels of periodicity or rates that together form a metrical hierarchy.
- Rhythm- related research has been an important area in Music Information Retrieval (MIR) since the discipline's earliest days. Audio tempo induction in particular has seen a lot of interest in the academic community. The aim of such systems is to extract a single-value descriptor that tells us something about the speed or pace of the music.
- MIR Music Information Retrieval
- Tempo is a minimal descriptor of rhythm content that represents one of the levels in the metrical hierarchy and, alone, does not provide information about broader structure.
- Induction of more than one metrical level is sometimes used as an intermediate step in beat tracking or onset detection methods.
- the metrical levels extracted in the aformentioned algorithms are labelled with respect to their musical or perceptive role.
- Gouyon et. al proposed a system differentiating between duple and triple meter implying the estimation of two metrical levels.
- Temperley and Sleator designed an algorithm to analyse meter and harmony in symbolic music data, which allowed the presence of more than three metrical levels.
- the present invention introduces a method to extract a complete metrical hierarchy from audio, representing it in a compact feature without applying semantic labels to any particular level.
- the extraction is performed using standard time-frequency analysis and a peak picking strategy that incorporates hierarchical structure constraints.
- the energy associated with each metrical level in our feature corresponds to the density of events of that rhythmic subdivision in the music signal.
- Appendix A of this document presents, in some detail, the mathematical algorithms employed by an example of the present invention to identify a consistent mathematical signature for the rhythmic structure of a given piece of audio content.
- the present invention requires that the pool of audio content from which recommendations are to be drawn first be analysed using the algorithmic techniques detailed for example in Appendix A.
- the present invention takes the following steps:
- such calculated metadata is calculated in advance and stored to assist in retrieving matching tracks which show similarities of rhythm, time signature(s), rhythmic complexity and/or cadence drift.
- tracks are deemed to be similar if one such calculated metadata metric matches exactly.
- tracks are deemed to be similar if one or more of such metrics lie within a defined range from that of the seed track.
- the extent of such a range is configurable.
- tracks are seemed to be similar if a defined number of such metadata metrics match exactly or lie within a defined range.
- the defined range is separately configurable for each such calculated metadata metric.
- those tracks may, in the preferred embodiment, be recommended to the end user whose seed track formed the basis for those recommendations .
- the list of recommended tracks is further narrowed down by reference to other criteria defined within the computer program or service within which the end user is listening to music.
- those tracks known (by inference or by direct selection by the end user or, in the preferred embodiment, by both methods) to be liked by a given end user have their calculated metadata metrics compared to attempt to divine a commonality of preferred music for that user.
- the aggregated metrics of those tracks is then, in the preferred embodiment, directly used to locate similar tracks to be recommended for that user, the search for similar tracks being performed in the same manner as disclosed for the "seed track" above, though utilising the aggregated metrics rather than those derived from the seed track to locate similar tracks.
- the calculated metadata metrics for a seed track is used to locate similar tracks for playback as a playlist of music which is related in style, such as by having similar rhythmic structures.
- the present invention may be further embodied to assist with both the provision of music recommendations and with other scenarios, including the following:
- Similarity-based computation based on rhythmic content and/or segmentation to identify audio tracks which are similar to one another, possibly for the purpose of recommending music based on an individual's existing taste in music.
- Measurement of rhythm similarity allows the present invention to be utilised to: o Recommend music which shares a similar rhythm to a given track o Assisted curation of music library (meaning that propositions are made by a computer to a human user). For example, this usage would enable the pre-selection of those tracks that would merge nicely for a given DJ. o Estimate the time stretching factor necessary to match two songs that don't exactly match according to the previous point,
- Reverse-engineer user or DJ signature generate a playlist that follows the pattern detected for a DJ or user.
- An automatic DJ software can be designed by determining the periodicities of beats of various musical tracks and then generating playlists comprising tracks which have similar rhythms.
- Intra-track similarity using result of segmentation algorithm, use the similarity measure disclosed above to assess similarity between parts of a track.
- rhythmic modulations using segmentation and similarity measures Generate "creative" recommendation. This would recommend a track that makes use of a rhythm structure that is a rhythmic modulation of the seed track, (this is a trick used by musicians, composers and some DJs) • Music classification. Based on the idea that different genres potentially have different rhythmic "signature”.
- rhythm may be matched to the users activity or bio feedback monitors
- a mid-level feature for metrical hierarchy Rhythm patterns in music are complex signals. In music that can be considered to have a sense of meter, the rhythm will have multiple related periodicities forming a hierarchical structure; the common musical terms for which include bars, beats and beat subdivisions.
- a mid-level feature that captures the weighting of salient pulse rates present in the music, which correspond to the most likely metrical hierarchy.
- Feature extraction is achieved by computing a novelty curve from the audio then processing this in time and frequency to produce a rhythmic spectrogram or rhythmogram. From this, we derive multiple possible candidates for a metrical hierarchy that fits the rhythmic content the best. We believe that this feature provides a meaningful description of the rhythm content without applying specific labels, such as tempo or tactus, to any particular level.
- the temporal occurrences of musical events (e.g. notes) in a musical piece can be, and often are, organised such that periodic patterns emerge. These can be decomposed into different levels of periodicity or rates that together form a metrical hierarchy.
- Rhythm- related research has been an important area in Music Information Retrieval (MIR) since the discipline's earliest days. Audio tempo induction in particular has seen a lot of interest in the community and is a long running track in the annual MIREX competition. The aim of such systems is to extract a single-value descriptor that tells us something about the speed or pace of the music.
- MIR Music Information Retrieval
- Tempo is a minimal descriptor of rhythm content that represents one of the levels in the metrical hierarchy and, alone, does not provide information about broader structure.
- Methods for extracting more than one metrical level from musical audio have also been proposed in the literature. Seppanen introduce a method to extract the 'beat and tatum levels' jointly and lapuri describes a method to extract the 'tatum, tactus and measure level'. Induction of more than one metrical level is sometimes used as an intermediate step in beat tracking or onset detection methods. With the exception of Degara's method, the metrical levels extracted in the aformentioned algorithms are labelled with respect to their musical or perceptive role. Gouyon et.
- Temperley and Sleator designed an algorithm to analyse meter and harmony in symbolic music data, which allowed the presence of more than three metrical levels.
- the metrical hierarchy that describes the rhythm structure of a piece of music may be represented as a tree as shown in Figure 1. Each horizontal level of nodes on the tree accounts for one metrical level (index ⁇ ⁇ [0> ⁇ ]) ; which is associated with a frequency,
- Metrical hierarchy is related in musical stave notation terms to the time signature and the note values used in a composition.
- time signature can be represented by Figure 1 (b), with metrical level i-1 being the bar level, level i the quarter note level (three quarter notes in one
- level i+ 1 being the eighth note level (quarter note divides into two eighth notes).
- metrical hierarchy can therefore be given as containing equivalent information to vector F .
- the symbol a is used to represent the fact that the frequency f
- Rhythmograms are a standard tool for rhythm analysis; a typical rhythmogram processing chain comprises two parts: estimation of a novelty curve capturing the sudden variations (e.g. transients) in the input audio signal x ⁇ and a time-frequency analysis that shows the periodicities pre-sent in .
- estimation of a novelty curve capturing the sudden variations (e.g. transients) in the input audio signal x ⁇
- a time-frequency analysis that shows the periodicities pre-sent in .
- rhythmograms obtained from standard computation of Fourier transform and autocorrelation function.
- Our input signal x ⁇ is mono audio sampled at 44.1kHz.
- We generate a simple novelty curve by full -wave rectifying taking the first derivative and half-wave rectifying the result:
- the autocorrelation rhythmogram is obtained with Grosche's toolbox also using a 12 second window length. For each window, the unbiased autocorrelation of ⁇ ( is computed. The lag scale is then mapped to BPM scale to generate rhythmogram
- FFT rhythmograms reveal harmonics of the periodicities present in the novelty curve whereas ACF rhythmograms reveal sub-harmonics of these same periodicities.
- Extraction of the feature from ⁇ is achieved using a peak-picking strategy constrained by the meter hierarchy model.
- a simple peak detection algorithm that detects a local maxima if an element is larger than both of its neighbours incorporating a threshold filter is employed (this was implemented using the standard Matlab function findpeaks).
- a threshold value of 0.085 was found to work well.
- each one of the metrical levels Ji is associated with a weight W > _
- Each hierarchy candidate is weighted by the sum of the weights of its metrical levels * .
- the hierarchy with the biggest weight ® is the one receiving the more energy from the musical signal, and is therefore considered as the hierarchy that best fits the data.
- a set of test rhythms were synthesised comprising all the permutations of groups of 3 or 4 events (24 patterns in total, which we labelled A-X), a selection of which is shown in Figure 3 tracks schema.
- An individual pattern would be synthesised as a 0.5 second audio sample using a cross-stick sound and each audio track is made up of that pattern repeating for 30 seconds.
- pattern Q is effectively an isochronous event train at 120 BPM.
- the corresponding spectrum in Figure 3 features only one peak at 120 BPM.
- Changing pattern Q to accented interpretation implies that the sequence of amplitudes of sonic events is modified from to 2 2 _ j n other words a rhythmic carrier frequency is excited, and therefore a peak appears at 360 BPM.
- Plain interpretation of pattern E excites two metrical levels: the carrier at 480 BPM because of the spacing between the two sounded notes and a 120 BPM level because of the repetition of the pattern.
- the high end of the extracted hierarchy for Steely Dan's 'Home at last' includes a triple ratio, consistent with the half-time 'Purdie shuffle' feel (easily scored in
- Vivaldi's 'Summer Presto' hierarchy also includes a triple ratio, but in the
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
There is provided a computer program product embodied on a non-transitory storage medium or on a cellular mobile telephone device or on another hardware device, the computer program product operable to automatically generate recommendations for digital media content, in which that system is adapted to automatically generate recommendations of digital media content for an end user by (a) analysing a first item or set of digital media content and its associated metadata to calculate metric-based metadata for that digital media content and (b) using that analysis to identify other digital media content which have a specifically-defined relationship with that first digital media content; and (c) provide, to the end user, recommendations of one or more items of that digital media content which were identified in the previous step.
Description
METHODS, SYSTEMS AND COMPUTER PROGRAM PRODUCTS FOR IDENTIFYING COMMONALITIES OF RHYTHM BETWEEN DISPARATE MUSICAL TRACKS AND USING THAT INFORMATION TO MAKE MUSIC RECOMMENDATIONS
BACKGROUND OF THE INVENTION
1. Field of the Invention The field of the invention relates to methods, systems and computer program products for identifying commonalities of rhythm between disparate musical tracks and using that information to make music recommendations.
2. Technical Background
A common problem in the field of music provision is the problem of which music tracks to recommend, both in terms of making explicit recommendations to individual consumers based on their existing musical taste and in terms of selecting a musical track to be played after any given musical track has concluded.
One approach to that problem which has been tried at various times is to analyse the current musical track (or the corpus of music which a given end user is known to enjoy) and identify common features with other tracks which fit well with the first track. However, the ways in which similarities between tracks are able to be automatically identified have historically been limited. The present invention teaches a method for more efficiently and reliably identifying commonalities of rhythm between disparate musical tracks and using that information to make recommendations accordingly.
SUMMARY OF THE INVENTION
The present invention provides a method for analysing audio media to extract metadata describing the rhythm, tempo and other descriptive metadata about the audio track and utilising that metadata to assist in cross -matching tracks, providing recommendations, perform additional Digital Signal Processing of tracks and for sundry other purposes as disclosed in this document.
The present invention teaches a method for efficiently and reliably identifying commonalities of rhythm between disparate musical tracks and using that information to make music recommendations accordingly.
Key Aspects of the Present Invention 1. A computer program product embodied on a non-transitory storage medium or on a cellular mobile telephone device or on another hardware device, the computer program product operable to automatically generate recommendations for digital media content, in which that system is adapted to automatically generate recommendations of digital media content for an end user by (a) analysing a first item or set of digital media content and its associated metadata to calculate metric-based metadata for that digital media content and (b) using that analysis to identify other digital media content which have a specifically-defined relationship with that first digital media content; and (c) provide, to the end user, recommendations of one or more items of that digital media content which were identified in the previous step.
2. The method of point 1 where a first digital media content is a track which is currently being listened to by an end user.
3. The method of point 1 where a first digital media content consists of a set of multiple tracks which are (i) analysed individually to calculate metric-based metadata and
(ii) that calculated metric-based metadata is aggregated and subsequently treated, for the purposes of subsequent steps in point 1, as if it derives from a single item of digital media content.
4. The method of point 3 where the set of multiple tracks comprises the set of digital media content files which a user enjoys, whether determined by observing that user's listening preferences or by the user specifically indicating his digital media content preferences or by any other means.
5. The method of any preceding point where the metric-based metadata comprises one or more of: the rhythmic value or values of a track, the time signature or time signatures of a track; the rhythmic simplicity or complexity of a track; the tempo or cadence drift of a track; the rhythm salience and rhythmicity; confidence values for any of the listed derived metrics; any other calculated metrics for a track.
6. The method of any preceding Claim where the specifically defined relationship with the first digital media content is that the identified other track or tracks share similar or identical calculated metric-based metadata with that first digital media content.
7. The method of point 6 where the calculated metric-based metadata for two or more tracks is defined as being similar if the two sets of metadata have values for a given metric which lie within a defined range from the metric-based metadata values calculated from a first item or set of digital media content.
8. The method of point 7 where the defined range may be defined separately for each metric-based metadata value which is utilised.
9. The method of any preceding point where the metric-based metadata is calculated in advance and stored in a database for comparison, search and retrieval purposes.
10. The method of any preceding point where the metric-based metadata is calculated as and when needed and discarded after use.
11. The method of any preceding point where the metric-based metadata is first calculated as and when needed and is then stored in a database for use subsequently.
12. The method of any preceding point where the one or more items of digital content recommended to the end user form a subset of those digital media content items identified as having similar or identical metric-based metadata values to the first item(s) of digital media content.
13. The method of point 12 where the subset of digital media content items is determined by filtering the list using additional criteria, including but not limited to the artist, album, genre, musical taste, current date, time of playback, date/time of most recent playback, user demographic data or by using any other additional criteria.
14. A computer based system adapted to perform the method of any preceding point.
BRIEF DESCRIPTION OF THE DRAWINGS
The above and other aspects of the invention will now be described, by way of example only, with reference to the following Figures, in which:
Figure 1 shows a Tree representation for metrical hierarchy, (a) A simple duple hierarchy
4
dividing the lower level into two groups of two (as with a 4 time signature), (b) A triple
3
duple hierarchy dividing the lower level into three groups of two (e.g. 4 time signature).
6
(c) A compound-duple hierarchy dividing the lower level into two groups of three (e.g. time signature).
Figure 2 shows an example of a special case with two metrical levels that can't coexist in the same hierarchy, hence requiring the creation of two parallel hierarchies: usage of triplets and eight notes at the same time (a.k.a 3:2 polyrh thm) represented by f J+2 and f, res ectively with 1 the quarter note level. Then
whereas
Figure 3 shows elementary patterns and the sequence of amplitudes associated to both their plain and accented interpretations.
Figure 4 shows Spectrum ^if) wjm extracted metrical hierarchy H , metrical levels rates F anc| relative weights vector W for (a) Billie Jean by Michael Jackson, (b) Home at last by Steely Dan and (c) Vivaldi's Violin Concerto in G minor (Summer), RV 315, movement III (Presto), as performed on the 1989 recording by Nigel Kennedy.
DETAILED DESCRIPTION Definitions For convenience, and to avoid needless repetition, the terms "music" and "media content" in this document are to be taken to encompass all "media content" which is in digital form or which it is possible to convert to digital form - including but not limited to books, magazines, newspapers and other periodicals, video in the form of digital video, motion pictures, television shows (as series, as seasons and as individual episodes), computer games and other interactive media, images (photographic or otherwise) and music.
Similarly, the term "track" indicates a specific item of media content, whether that be a song, a television show, an eBook or portion thereof, a computer game or any other discreet item of media content.
The terms "playlist", "timeline" and "album" are used interchangeably to indicate collections of "tracks" and/or interstitials which have been conjoined together such that they may be treated as a single entity for the purposes of analysis or recommendation.
The terms "digital media catalogue", "digital music catalogue", "media catalogue" and "catalogue" are used interchangeably to indicate a collection of tracks and/ or albums to which a user may be allowed access for listening purposes. The digital media catalogue may aggregate both digital media files and their associated metadata or, in another example embodiment, the digital media and metadata may be delivered from multiple such catalogues. There is no implication that only one such catalogue exists, and the term encompasses access to multiple separate catalogues simultaneously, whether consecutively, concurrently or by aggregation. The actual catalogue utilised by any given operation may be fixed or may vary over time and/ or according to the location or access rights of a particular device or end-user.
The abbreviation "DRM" is used to refer to a "Digital Rights Management" system or mechanism used to grant access rights to a digital media file.
The verbs "to listen", "to view" and "to play" are to be taken as encompassing any interaction between a human and media content, whether that be listening to audio content, watching video or image content, reading books or other textual content, playing a computer game, interacting with interactive media content or some combination of such activities. In addition, the playing of digital media content is to be further construed as occurring where some or all of one digital media content file is incorporated within another digital media content file such that playing the latter would necessitate the former also being played. The terms "user", "consumer", "end user" and "individual" are used interchangeably to refer to the person, or group of people making use of the facilities provided by the interface. In all cases, the masculine includes the feminine and vice versa.
The terms "device" and "media player" are used interchangeably to refer to any computational device which is capable of playing digital media content, including but not limited to MP3 players, television sets, home entertainment system, home computer systems, mobile computing devices, games consoles, handheld games consoles, IVEs or other vehicular-based media players or any other applicable device or software media player on such a device. Something essentially capable of playback of media.
The term "DSP" ("Digital Signal Processing") refers to any computational processing of digital media content in order to extract additional metadata from that content. Such calculated metadata may take a variety of forms, including deriving the tempo of a musical track or identifying one or more spots within the digital media file which are gauged to be representative of that content as a whole.
The term "hook" is used to refer to one or more portions of a digital media file which have been identified, whether via DSP or manually or by some other method, as being representative of the content as a whole. For example, a movie trailer consists of a series of one or more "hooks" from the movie while particularly apposite riffs or lines from a musical track serve a similar identifying purpose.
The terms "partial playback" and "extract" are used interchangeably to refer to the playback of an item of digital media content where that item is not played back in its entirety or is incorporated into another item of digital media content. The abbreviation "BPM" is used to denote the count of the number of Beats Per Minute in an audio track.
The abbreviation "FFT" is used for the mathematical term "Fast Fourier Transform". The abbreviation "MIR" is used to denote "Music Information Retrieval", a field of study whereby audio signals are analysed to extract metadata such as, and most relevantly to the present invention, the extraction of rhythm metadata from audio tracks.
The terms "DJ" and "Disk Jockey" are used interchangeably to refer to individuals who compile and/ or play collections, or "mix tapes", of musical tracks.
Tempo
Tempo estimation has been an important aspect of Music Information Retrieval (MIR) research since its inception. A large body of research has been devoted to the task. Tempo is commonly considered as a fundamental property of the music, which certainly explains why it has been one of the first MIR tasks to be investigated.
Tempo is often associated to our perception of the speed or pace of the music or, in other words, how fast we feel the music is. Its relevance for music databases navigation, recommendation systems or automatic curation is then obvious.
However, the definition of the term "tempo" itself is not consistent in the related MIR literature thus leaving room for ambiguity and fuzziness in the discussions.
Metronomic indications vs. tempo
Published academic papers almost exclusively represent tempo as a BPM (Beats per Minute) value, which is effectively a frequency measurement. The implicit assumption
made there is that tempo can be represented by a single rate (frequency). Given that tempo is commonly associated with speed or pace of music, that is to say our sense of how fast the music is is governed by one rate. A nuance between metronomic indication and tempo need to be made here as are generally represented in the form of a BPM value. A metronomic indication is written on a music score and associates a note value with an absolute frequency measurement, thereby providing an absolute point of reference for speed of execution - usually using a metronome. This has several implications. First, a metronomic indication makes sense only in regard to the score it is attached to. Secondly, the "correct" metronomic indication for a given score is not necessarily unique. Numerous examples of this can be found in music history, Chopin being an example of a composer who used different notations on different editions of his scores. A metronomic indication is relevant for performers, not necessarily for listeners. The score is a cue to the performers so that they can deliver the musical feel intended by the composer (including the sense of pace) but it does not mean a single element of the score (the metronomic indication) bears this information. In that sense, Italian tempo indications such as presto, allegro, lento provide information which would be a lot more meaningful to listeners than just a metronomic indication since they describe what the music should sound like rather than the prescription of a technical performance element. Of course music notation is subject to conventions that influence the performance of musicians. For example, a cut time score usually bears a connotation for fast music. But the point being made here is that although metronomic indication and sense of speed or pace are not uncorrelated, there isn't a strict equivalence between them.
Tactus vs. tempo
With the design of tempo estimation algorithms from audio comes the need to evaluate them. This is commonly achieved by comparing the output of the algorithm with some ground truth data. There are two common possibility for ground truth tempo known as "notated tempo" and "perceived tempo".
The "notated tempo" is obtained from the score and effectively refers to the metronomic indication. The access to, or even the existence of, a music score with a metronomic indication is not easy, and is furthermore subject to copyright law. For that reason, unless working on symbolic data, this source of ground truth tends not to be commonly used. The "perceived tempo" ground truth data is typically collected by asking people to tap along a music piece, record the hits and measure the tapping rate. This is what the ground truth data used for the MIREX audio tempo estimation task consists of.
In more rigorous terms, the rate at which listeners tap along to music is called 'tactus'. That type of evaluation has become a standard in tempo estimation evaluation and is based on the implicit assumption that tactus equals tempo.
However, that assumption is challengeable. Going beyond tempo
The present invention provides a descriptor that is more closely related to the temporal feel of music than a single rate (whether it is called tempo or tactus). Such a descriptor has direct applications in the context of today's music distribution and consumption such as content-based recommendation for streaming and download services, automatic radio, curation assistance etc.
The metrical structure appears to play an important role in perception of speed: differences in surface rhythm and metrical structure do interfere with judgments of tempo, even if two passages have the same beat rate. Consequently, an important focus of the present invention is on metrical structure extraction from audio. It is also safe to assume that the metrical structure influences more that just the sense of speed. For instance, music in a compound meter feels different from a simple meter.
Detailed Description
Rhythm patterns in music are complex signals. In music which can be considered to have a sense of meter, the rhythm will have multiple related periodicities forming a
hierarchical structure; the common musical terms for which include bars, beats and beat subdivisions.
The present invention introduces a mid-level feature which captures the weighting of salient pulse rates present in the music, which correspond to the most likely metrical hierarchy. Feature extraction is achieved by computing a novelty curve from the audio then processing this in time and frequency to produce a rhythmic spectrogram or rhythmogram. From this, we derive multiple possible candidates for a metrical hierarchy that fits the rhythmic content the best. That feature provides a meaningful description of the rhythm content without applying specific labels, such as tempo or tactus, to any particular level.
Overview The temporal occurrences of musical events (e.g. notes) in a musical piece can be, and often are, organised such that periodic patterns emerge. These can be decomposed into different levels of periodicity or rates that together form a metrical hierarchy. Rhythm- related research has been an important area in Music Information Retrieval (MIR) since the discipline's earliest days. Audio tempo induction in particular has seen a lot of interest in the academic community. The aim of such systems is to extract a single-value descriptor that tells us something about the speed or pace of the music.
Tempo is a minimal descriptor of rhythm content that represents one of the levels in the metrical hierarchy and, alone, does not provide information about broader structure.
Induction of more than one metrical level is sometimes used as an intermediate step in beat tracking or onset detection methods. With the exception of Degara's method, the metrical levels extracted in the aformentioned algorithms are labelled with respect to their musical or perceptive role. Gouyon et. al proposed a system differentiating between duple and triple meter implying the estimation of two metrical levels. Further back, Temperley and Sleator designed an algorithm to analyse meter and harmony in symbolic music data, which allowed the presence of more than three metrical levels.
The present invention introduces a method to extract a complete metrical hierarchy from audio, representing it in a compact feature without applying semantic labels to any particular level. The extraction is performed using standard time-frequency analysis and a peak picking strategy that incorporates hierarchical structure constraints. The energy associated with each metrical level in our feature corresponds to the density of events of that rhythmic subdivision in the music signal.
Appendix A of this document presents, in some detail, the mathematical algorithms employed by an example of the present invention to identify a consistent mathematical signature for the rhythmic structure of a given piece of audio content.
Track Similarity
The present invention requires that the pool of audio content from which recommendations are to be drawn first be analysed using the algorithmic techniques detailed for example in Appendix A.
Once that step is completed, whether carried out on the fly or, in the preferred embodiment, carried out in advance, one or more of the following items of metadata will be available for each musical track which is so analysed:
• A metric which represents the rhythmic value or values of that piece of audio
• A metric which represents the time signature or signatures of that piece of audio
• A metric which represents the rhythmic simplicity or complexity of that piece of audio
· Values to establish the tempo or cadence drift of that piece of audio
• A metric representing the rhythm salience and rhythmicity of that piece of audio
• A confidence value for the extracted rhythm metric.
In the preferred embodiment of the present invention, all such values are calculated in advance for all musical tracks from which recommendations are to be drawn. The results of such calculations are, in the preferred embodiment, stored in a database via which those values may be searched.
When recommending additional music based on a given musical track (the "seed track"), the present invention takes the following steps:
1. Calculate (or retrieve pre-calculated values from a database) the desired metadata for the given track;
2. Locate other musical tracks which share similar or identical metrics within the database of such calculated metadata;
3. Locate other musical tracks whose metrics have a specific relationship to the seed. This relationship is not necessarily similarity and can be specified in the embodiment of the present invention. In a further example embodiment of the present invention, multiple types of inter- track compatibility (defined by the given relationship, and not necessarily equal to similarity) may be employed.
In the preferred embodiment of the present invention, such calculated metadata is calculated in advance and stored to assist in retrieving matching tracks which show similarities of rhythm, time signature(s), rhythmic complexity and/or cadence drift. In another embodiment of the present invention, tracks are deemed to be similar if one such calculated metadata metric matches exactly. In other example embodiment, tracks are deemed to be similar if one or more of such metrics lie within a defined range from that of the seed track. In the preferred embodiment, the extent of such a range is configurable. In the preferred embodiment, tracks are seemed to be similar if a defined number of such metadata metrics match exactly or lie within a defined range. In the preferred embodiment, the defined range is separately configurable for each such calculated metadata metric.
Uses in Recommendations
When one or more tracks are identified, using the method disclosed above, as being "similar" to the seed track, those tracks may, in the preferred embodiment, be recommended to the end user whose seed track formed the basis for those recommendations .
In the preferred embodiment, the list of recommended tracks is further narrowed down by reference to other criteria defined within the computer program or service within which the end user is listening to music. In another example embodiment, those tracks known (by inference or by direct selection by the end user or, in the preferred embodiment, by both methods) to be liked by a given end user have their calculated metadata metrics compared to attempt to divine a commonality of preferred music for that user. The aggregated metrics of those tracks is then, in the preferred embodiment, directly used to locate similar tracks to be recommended for that user, the search for similar tracks being performed in the same manner as disclosed for the "seed track" above, though utilising the aggregated metrics rather than those derived from the seed track to locate similar tracks.
In a further example embodiment, the calculated metadata metrics for a seed track is used to locate similar tracks for playback as a playlist of music which is related in style, such as by having similar rhythmic structures.
Additional Use Cases In addition to the preferred embodiment of the present invention disclosed above, the present invention may be further embodied to assist with both the provision of music recommendations and with other scenarios, including the following:
• Segmentation of an audio track based on the varying rhythmic values detected in that piece of audio
· Similarity-based computation based on rhythmic content and/or segmentation, to identify audio tracks which are similar to one another, possibly for the purpose of recommending music based on an individual's existing taste in music.
Calculate difference values for pitch and tempo changes for beat matching of different audio tracks
Identification of machine-generated tempo clocks vs live clocks
Identification of hook points in a given audio track
Identification of tempo changes in a given audio track
Use of the start and end tempos for sequencing of media
A combination of all elements disclosed herein alongside existing metadata to improve or expand on recommendations, sequencers or AI media suggestions i.e. using time of day, genre, to identify the rhythmical style of music via speed, frequency or complexity in order to recommend music at different times of day. Use the present invention to inform a beat tracking system, so that the periodicities of the beats that should be found are already known.
Measurement of rhythm similarity allows the present invention to be utilised to: o Recommend music which shares a similar rhythm to a given track o Assisted curation of music library (meaning that propositions are made by a computer to a human user). For example, this usage would enable the pre-selection of those tracks that would merge nicely for a given DJ. o Estimate the time stretching factor necessary to match two songs that don't exactly match according to the previous point,
o Evaluation of listeners tastes: do they always listen to something that has the same signature in terms of feature values (in other word is the statistical inter track distance low?) ?
o Similarly, are there patterns in mix tapes, set lists or playlists made by musicians, DJs, etc. . . This would possibly give the ability to spot the "signature" of DJs.
Reverse-engineer user or DJ signature: generate a playlist that follows the pattern detected for a DJ or user.
An automatic DJ software can be designed by determining the periodicities of beats of various musical tracks and then generating playlists comprising tracks which have similar rhythms.
Segmentation of a piece
Intra-track similarity: using result of segmentation algorithm, use the similarity measure disclosed above to assess similarity between parts of a track.
Extraction of a speed feeling value. This would answer the question: how fast does that music piece feels?
Detection of rhythmic modulations using segmentation and similarity measures Generate "creative" recommendation. This would recommend a track that makes use of a rhythm structure that is a rhythmic modulation of the seed track, (this is a trick used by musicians, composers and some DJs)
• Music classification. Based on the idea that different genres potentially have different rhythmic "signature".
• Estimation of the presence/absence of drums/percussions in a music piece.
• Entrainment for biomusicology— rhythm may be matched to the users activity or bio feedback monitors
• For use in biomusicology to affect a user or set of users in a desired way
• In car/vehicle matching of various aspects of things like speed, gear changes, etc
• For journeys where rhythm can be applied to an ETA
• For musical sequencing where tempo, pulse or rhythm is part of the track to track journey or tempo curve over time and multiple songs/playlists/radio stations
• For use in Radio sequencers or schedulers
• For use in track identification or matching
• For use in searching for musical pieces that could fit into a music production in a studio or other similar creative environment
• For use as part of a sub set filing system that locates music for production by matching rhythm
• For use in machine learning
• For use in synchronisation of musical or audio pieces
· Identification of sub set rhythmic patterns or harmonically diverse (or simple) pieces of music
• Identification of non audible rhythmic patterns for use in data analyses of high resolution audio or media
• Scientific data analyses of any rhythmic pattern in data sets (beyond audio) · Identification of 'harding' or pulsing of images and light in video/visual mediums
• Use in live audio situations for teaching or performance
• Use for multiplexing and de-multiplexing periodic signals
• navigation of large data sets, libraries and catalogues Note
It is to be understood that the above-referenced arrangements are only illustrative of the application for the principles of the present invention. Numerous modifications and
alternative arrangements can be devised without departing from the spirit and scope of the present invention. While the present invention has been shown in the drawings and fully described above with particularity and detail in connection with what is presently deemed to be the most practical and preferred example(s) of the invention, it will be apparent to those of ordinary skill in the art that numerous modifications can be made without departing from the principles and concepts of the invention as set forth herein.
APPENDIX A
A mid-level feature for metrical hierarchy Rhythm patterns in music are complex signals. In music that can be considered to have a sense of meter, the rhythm will have multiple related periodicities forming a hierarchical structure; the common musical terms for which include bars, beats and beat subdivisions. In this Appendix, we introduce a mid-level feature that captures the weighting of salient pulse rates present in the music, which correspond to the most likely metrical hierarchy. Feature extraction is achieved by computing a novelty curve from the audio then processing this in time and frequency to produce a rhythmic spectrogram or rhythmogram. From this, we derive multiple possible candidates for a metrical hierarchy that fits the rhythmic content the best. We believe that this feature provides a meaningful description of the rhythm content without applying specific labels, such as tempo or tactus, to any particular level.
1. Introduction
The temporal occurrences of musical events (e.g. notes) in a musical piece can be, and often are, organised such that periodic patterns emerge. These can be decomposed into different levels of periodicity or rates that together form a metrical hierarchy. Rhythm- related research has been an important area in Music Information Retrieval (MIR) since the discipline's earliest days. Audio tempo induction in particular has seen a lot of interest in the community and is a long running track in the annual MIREX competition. The aim of such systems is to extract a single-value descriptor that tells us something about the speed or pace of the music.
Tempo is a minimal descriptor of rhythm content that represents one of the levels in the metrical hierarchy and, alone, does not provide information about broader structure. Methods for extracting more than one metrical level from musical audio have also been proposed in the literature. Seppanen introduce a method to extract the 'beat and tatum levels' jointly and lapuri describes a method to extract the 'tatum, tactus and measure level'.
Induction of more than one metrical level is sometimes used as an intermediate step in beat tracking or onset detection methods. With the exception of Degara's method, the metrical levels extracted in the aformentioned algorithms are labelled with respect to their musical or perceptive role. Gouyon et. al proposed a system differentiating between duple and triple meter implying the estimation of two metrical levels. Further back, Temperley and Sleator designed an algorithm to analyse meter and harmony in symbolic music data, which allowed the presence of more than three metrical levels.
In this Appendix we introduce a method to extract a complete metrical hierarchy from audio, representing it in a compact feature without applying semantic labels to any particular level. The extraction is performed using standard time-frequency analysis and a peak picking strategy that incorporates hierarchical structure constraints. The energy associated with each metrical level in our feature corresponds to the density of events of that rhythmic subdivision in the music signal.
2. Formalising metrical hierarchy
The metrical hierarchy that describes the rhythm structure of a piece of music may be represented as a tree as shown in Figure 1. Each horizontal level of nodes on the tree accounts for one metrical level (index ζ ^[0> ^]); which is associated with a frequency,
f
or pulse rate J l measured in BPM (Beats Per Minute). The number of metrical levels necessary to represent the rhythm hierarchy of a piece of music is therefore N + 1 . These rates can be grouped in ascending order in a vector ^ (fo> fi> -> N Hierarchical relationships are defined by the number Λ Qf child nodes each level generates. This
A - f>
implies that J>~1 . As a consequence, the full hierarchical structure can be described as a sequence of frequency ratios ^ ^1' .
Metrical hierarchy is related in musical stave notation terms to the time signature and the note values used in a composition. As an example, a musical piece using eighth notes in a 3
4 time signature can be represented by Figure 1 (b), with metrical level i-1 being the bar
level, level i the quarter note level (three quarter notes in one
3
4 bar) and level i+ 1 being the eighth note level (quarter note divides into two eighth notes).
With the ratios defined in Λ ? specification of frequency for a single metrical level allows the frequencies of all other levels to be calculated. A simple expression of the
H = f ειΛ
metrical hierarchy can therefore be given as containing equivalent information to vector F . (The symbol a is used to represent the fact that the frequency f
The feature we introduce in this Appendix can be fully represented by F Qr H coupled with a vector of weights W eg. as described below. This provides a rich and compact description of the rhythmic content by associating a hierarchical structure with a relative weighting of the metrical levels.
3 Time-frequency analysis Rhythmograms are a standard tool for rhythm analysis; a typical rhythmogram processing chain comprises two parts: estimation of a novelty curve capturing the sudden variations (e.g. transients) in the input audio signal x^ and a time-frequency analysis that shows the periodicities pre-sent in . To extract our rhythm feature, we exploit rhythmograms obtained from standard computation of Fourier transform and autocorrelation function. Our input signal x^ is mono audio sampled at 44.1kHz. We generate a simple novelty curve by full -wave rectifying
taking the first derivative and half-wave rectifying the result:
Δ(ί) =\ ^- (\ x(t) I)
dt (i)
A framed FFT of is computed (using a 12 second Hamming window with 97% overlap) resulting in the rhythmogram f) .
The autocorrelation rhythmogram is obtained with Grosche's toolbox also using a 12 second window length. For each window, the unbiased autocorrelation of Δ( is computed. The lag scale is then mapped to BPM scale to generate rhythmogram
For this Appendix we calculate our metrical hierarchy feature across a whole song so we average the rhythmograms ^F (^ /) and ^ (t ) over time. To obtain the average rhythmic sp ogram we calculate
where X denotes which rhythmogram or ^4 ) is being averaged.
FFT rhythmograms reveal harmonics of the periodicities present in the novelty curve whereas ACF rhythmograms reveal sub-harmonics of these same periodicities.
Multiplying the spectra ΩΛ/) and ΩΛ (/) therefore allows us to keep only the common periodicities, the most salient of which correlate with the metrical rates used in the music. Following that rationale, to produce a final rhythmic spectrum vector ^ we calculate the Hadamard product (an element by element multiplication denoted as °
) of the spectra ΩΛ/) and ΩΛ (/) and normalise the result:
(QA(f) o QF(f))
S(f)
max (Q, (/) o Q (/))
/ (3)
4 Hierarchy feature extraction
Extraction of the feature from ^ is achieved using a peak-picking strategy constrained by the meter hierarchy model. A simple peak detection algorithm that detects a local maxima if an element is larger than both of its neighbours incorporating a
threshold filter is employed (this was implemented using the standard Matlab function findpeaks). A threshold value of 0.085 was found to work well.
The abcissa of biggest peak in ^ , frequency f max ^ js usecJ as a point of reference to extract the metrical hierarchy. It represents the rhtyhmic rate containing the most energy in the audio signal and is therefore assumed to represent an important metrical level. If
f JflCtX
the peak at is part of the hierarchical structure, this directly implies that the rates
f f JflCtX
of all other metrical levels 1 are related to by integer ratios. In order to implement this constraint, the abscissa ^ 1 of the ^ ^ peak in S(f^ is compared to f JflCtX
and is considered as a candidate level if, and only if, it satisfies one of the following conditions
otherwise it is rejected.
By finding all the peaks that are integer ratios of we cannot guarantee that they form a hierarchy consistent with the model. Moreover a single hierarchy might not be enough to describe a rhythmic structure (polyrhythmic music may imply multiple competing hierarchies for example). A further constraint therefore needs to be added to build a consistent hierarchy: the rate of each metrical level and its immediate neighbour must be related by an integer ratio. f JflCtX
As before, the biggest peak (at ) is taken as a reference as it must be part of the hierarchical structure. Starting with f max ^ iterative comparison of metrical levels is performed upwards (comparison with metrical levels with higher rates) and downwards
(comparison with metrical levels with lower rates). Moving upwards, a metrical level f +1 f
is accepted as part of the hierarchy if ■> is an integer number, otherwise it is rejected.
In this case 1 will then be compared to the next rate J+2 and so on until an integer ratio is found or the peaks are exhausted. f j+1 f j + 2 f j + 2
A special case occurs when f■> and f 1 are integer ratios but f 1+1 i .s not. This means that the metrical level associated to 1 could equally be subdivided in levels +1 or J+2 whereas these two levels can't coexist in the same hierarchy (as shown in Figure 2). In such a situation, two parallel hierarchy candidates are generated; one using f J+l (the duple meter in the example) and the other using f J+2 (the triple meter in the example). At this stage, hierarchy candidates have been generated, and are equally represented by their vector F Qr H .
Finally, for each hierarch candidate, each one of the metrical levels Ji is associated with a weight W>
_ Each hierarchy candidate is weighted by the sum of the weights of its metrical levels * . The hierarchy with the biggest weight ® is the one receiving the more energy from the musical signal, and is therefore considered as the hierarchy that best fits the data.
5 Examples 5.1 Synthesised test tracks
A set of test rhythms were synthesised comprising all the permutations of groups of 3 or 4 events (24 patterns in total, which we labelled A-X), a selection of which is shown in Figure 3 tracks schema. An individual pattern would be synthesised as a 0.5 second audio sample using a cross-stick sound and each audio track is made up of that pattern repeating for 30 seconds.
[plainE] [plainK] [plainQ]
[accentsE] [accents K] [accents Q]
Spectrums ^ with extracted hierarchy H ; metrical levels rates F and relative weights vector W , for both plain and accented interpretation of patterns E, K and Q.
Two interpretations for each pattern were produced: Firstly, a plain interpretation for which black circles represent sounded notes (amplitude =1), and dashes represent rests (amplitude =0). Secondly, an accented interpretation for which circles are interpreted as accented notes (amplitude =1) whereas dashes are interpreted as unaccented notes (in this
1_
case amplitude= 2 ).
[Pattern E
Plain (^AO)
Accented 2 2 ] (Q.2,0) circle (0.2cm); (1,0) circle (0.2cm); [thick] (1.6,0) - (2,0); [thick] (2.4,0) - (2.8,0); [Pattern Q
Plain <1.0,0>
Accented 2 2 ] (Q.2,0) circle (0.2cm); [thick] (0.8,0) - (1.2,0); [thick] (1.6,0) - (2,0); [Pattern K
Plain <Μ θ
Cu
Accented ,i 2> ] (Q.2,0) circle (0.2cm); (1,0) circle (0.2cm); (1.8,0) circle (0.2cm);
[thick] (2.4,0) - (2.8,0); [Pattern T
Plain <U >°>
Accented Cu 2> ] (Q.2,0) circle (0.2cm); (1,0) circle (0.2cm); [thick] (1.6,0) - (2,0);
The plain interpretation of pattern Q is effectively an isochronous event train at 120 BPM. As expected, the corresponding spectrum in Figure 3 features only one peak at 120 BPM. Changing pattern Q to accented interpretation implies that the sequence of amplitudes of sonic events is modified from to 2 2 _ jn other words a rhythmic carrier frequency is excited, and therefore a peak appears at 360 BPM. Plain
interpretation of pattern E excites two metrical levels: the carrier at 480 BPM because of the spacing between the two sounded notes and a 120 BPM level because of the repetition of the pattern. Plain interpretation of pattern K excites three metrical levels: the 480 BPM carrier level because of the spacing between first and second and second and third note, the 120 BPM level due to the pattern repetition and finally 240 BPM, the presence of which is explained by notes in first and third position being sounded.
Using accented interpretation effectively means that there are more notes played at the carrier metrical level, and therefore more energy present at that frequency. This fact is captured in W where the relative weight of carrier level is bigger. This demonstrates that the feature not only captures which metrical levels are used but also how much they are used.
5.2 Commercial tracks
We now apply the same analysis to commercial music recordings with results shown in Figure 4. The extracted hierarchy for Michael Jackson's 'Billie Jean' comprises only duple ratios. It is consistent with the fact that this track has a straight feel, easily scored in
4
4 . In contrast, the high end of the extracted hierarchy for Steely Dan's 'Home at last' includes a triple ratio, consistent with the half-time 'Purdie shuffle' feel (easily scored in
12
8 ). Finally, Vivaldi's 'Summer Presto' hierarchy also includes a triple ratio, but in the
3 low-end of the hierarchy this time. It is consistent with the scored meter of 4 . In fact, assuming the metrical level at 149 BPM is the quarter note, the other extracted levels
3
would then correspond to: two bars, one 4 bar, quarter note, eighth notes and sixteenth notes. Here again the presence of the 607BPM metrical level and its relatively high weighting correspond to the heavy use of sixteenth notes made throughout the piece. Similarly, in Billie Jean, considering the 117 BPM metrical level as the quarter note, the composition is heavily based on eighth notes, in accordance with the predominant importance of the 234 BPM metrical level.
6 Conclusion
In this Appendix we have introduced a mid-level descriptor that captures rhythm structure information from audio. The examples shown here have allowed us to demonstrate that a feature of low dimensionality can jointly capture metrical hierarchy and the relative weighting of each metrical level along with a quantitative speed reference. It therefore provides a compact representation of musically meaningful information about the rhythmic content of a music piece.
Note
It is to be understood that the above-referenced arrangements are only illustrative of the application for the principles of the present invention. Numerous modifications and alternative arrangements can be devised without departing from the spirit and scope of the present invention. While the present invention has been shown in the drawings and fully described above with particularity and detail in connection with what is presently deemed to be the most practical and preferred example(s) of the invention, it will be apparent to those of ordinary skill in the art that numerous modifications can be made without departing from the principles and concepts of the invention as set forth herein.
Claims
1. A computer program product embodied on a non-transitory storage medium or on a cellular mobile telephone device or on another hardware device, the computer program product operable to automatically generate recommendations for digital media content, in which that system is adapted to automatically generate recommendations of digital media content for an end user by (a) analysing a first item or set of digital media content and its associated metadata to calculate metric-based metadata for that digital media content and (b) using that analysis to identify other digital media content which have a specifically-defined relationship with that first digital media content; and (c) provide, to the end user, recommendations of one or more items of that digital media content which were identified in the previous step.
2. The method of preceding Claim 1 where a first digital media content is a track which is currently being listened to by an end user.
3. The method of preceding Claim 1 where a first digital media content consists of a set of multiple tracks which are (i) analysed individually to calculate metric-based metadata and (ii) that calculated metric-based metadata is aggregated and subsequently treated, for the purposes of subsequent steps in claim 1, as if it derives from a single item of digital media content.
4. The method of preceding Claim 3 where the set of multiple tracks comprises the set of digital media content files which a user enjoys, whether determined by observing that user's listening preferences or by the user specifically indicating his digital media content preferences or by any other means.
5. The method of any preceding Claim where the metric-based metadata comprises one or more of: the rhythmic value or values of a track, the time signature or time signatures of a track; the rhythmic simplicity or complexity of a track; the tempo or cadence drift of a track; the rhythm salience and rhythmicity; confidence values for any of the listed derived metrics; any other calculated metrics for a track.
6. The method of any preceding Claim where the specifically defined relationship with the first digital media content is that the identified other track or tracks share similar or identical calculated metric-based metadata with that first digital media content.
7. The method of preceding Claim 6 where the calculated metric-based metadata for two or more tracks is defined as being similar if the two sets of metadata have values for a given metric which lie within a defined range from the metric-based metadata values calculated from a first item or set of digital media content.
8. The method of preceding Claim 7 where the defined range may be defined separately for each metric-based metadata value which is utilised.
9. The method of any preceding Claim where the metric-based metadata is calculated in advance and stored in a database for comparison, search and retrieval purposes.
10. The method of any preceding Claim where the metric-based metadata is calculated as and when needed and discarded after use.
11. The method of any preceding Claim where the metric-based metadata is first calculated as and when needed and is then stored in a database for use subsequently.
12. The method of any preceding Claim where the one or more items of digital content recommended to the end user form a subset of those digital media content items identified as having similar or identical metric-based metadata values to the first item(s) of digital media content.
13. The method of preceding Claim 12 where the subset of digital media content items is determined by filtering the list using additional criteria, including but not limited to the artist, album, genre, musical taste, current date, time of playback, date/time of most recent playback, user demographic data or by using any other additional criteria.
14. A computer based system adapted to perform the method of any preceding Claim.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB1408257.2 | 2014-05-09 | ||
GBGB1408257.2A GB201408257D0 (en) | 2014-05-09 | 2014-05-09 | Rhythm metrics |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015170126A1 true WO2015170126A1 (en) | 2015-11-12 |
Family
ID=51032518
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/GB2015/051380 WO2015170126A1 (en) | 2014-05-09 | 2015-05-11 | Methods, systems and computer program products for identifying commonalities of rhythm between disparate musical tracks and using that information to make music recommendations |
Country Status (2)
Country | Link |
---|---|
GB (1) | GB201408257D0 (en) |
WO (1) | WO2015170126A1 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10672371B2 (en) | 2015-09-29 | 2020-06-02 | Amper Music, Inc. | Method of and system for spotting digital media objects and event markers using musical experience descriptors to characterize digital music to be automatically composed and generated by an automated music composition and generation engine |
US10854180B2 (en) | 2015-09-29 | 2020-12-01 | Amper Music, Inc. | Method of and system for controlling the qualities of musical energy embodied in and expressed by digital music to be automatically composed and generated by an automated music composition and generation engine |
US10964299B1 (en) | 2019-10-15 | 2021-03-30 | Shutterstock, Inc. | Method of and system for automatically generating digital performances of music compositions using notes selected from virtual musical instruments based on the music-theoretic states of the music compositions |
US11024275B2 (en) | 2019-10-15 | 2021-06-01 | Shutterstock, Inc. | Method of digitally performing a music composition using virtual musical instruments having performance logic executing within a virtual musical instrument (VMI) library management system |
US11037538B2 (en) | 2019-10-15 | 2021-06-15 | Shutterstock, Inc. | Method of and system for automated musical arrangement and musical instrument performance style transformation supported within an automated music performance system |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110295843A1 (en) * | 2010-05-26 | 2011-12-01 | Apple Inc. | Dynamic generation of contextually aware playlists |
US20130031162A1 (en) * | 2011-07-29 | 2013-01-31 | Myxer, Inc. | Systems and methods for media selection based on social metadata |
WO2014002064A1 (en) * | 2012-06-29 | 2014-01-03 | Ecole Polytechnique Federale De Lausanne (Epfl) | System and method for media library navigation and recommendation |
-
2014
- 2014-05-09 GB GBGB1408257.2A patent/GB201408257D0/en not_active Ceased
-
2015
- 2015-05-11 WO PCT/GB2015/051380 patent/WO2015170126A1/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110295843A1 (en) * | 2010-05-26 | 2011-12-01 | Apple Inc. | Dynamic generation of contextually aware playlists |
US20130031162A1 (en) * | 2011-07-29 | 2013-01-31 | Myxer, Inc. | Systems and methods for media selection based on social metadata |
WO2014002064A1 (en) * | 2012-06-29 | 2014-01-03 | Ecole Polytechnique Federale De Lausanne (Epfl) | System and method for media library navigation and recommendation |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11037539B2 (en) | 2015-09-29 | 2021-06-15 | Shutterstock, Inc. | Autonomous music composition and performance system employing real-time analysis of a musical performance to automatically compose and perform music to accompany the musical performance |
US11037541B2 (en) | 2015-09-29 | 2021-06-15 | Shutterstock, Inc. | Method of composing a piece of digital music using musical experience descriptors to indicate what, when and how musical events should appear in the piece of digital music automatically composed and generated by an automated music composition and generation system |
US10672371B2 (en) | 2015-09-29 | 2020-06-02 | Amper Music, Inc. | Method of and system for spotting digital media objects and event markers using musical experience descriptors to characterize digital music to be automatically composed and generated by an automated music composition and generation engine |
US11037540B2 (en) | 2015-09-29 | 2021-06-15 | Shutterstock, Inc. | Automated music composition and generation systems, engines and methods employing parameter mapping configurations to enable automated music composition and generation |
US11017750B2 (en) | 2015-09-29 | 2021-05-25 | Shutterstock, Inc. | Method of automatically confirming the uniqueness of digital pieces of music produced by an automated music composition and generation system while satisfying the creative intentions of system users |
US10854180B2 (en) | 2015-09-29 | 2020-12-01 | Amper Music, Inc. | Method of and system for controlling the qualities of musical energy embodied in and expressed by digital music to be automatically composed and generated by an automated music composition and generation engine |
US11030984B2 (en) | 2015-09-29 | 2021-06-08 | Shutterstock, Inc. | Method of scoring digital media objects using musical experience descriptors to indicate what, where and when musical events should appear in pieces of digital music automatically composed and generated by an automated music composition and generation system |
US11657787B2 (en) | 2015-09-29 | 2023-05-23 | Shutterstock, Inc. | Method of and system for automatically generating music compositions and productions using lyrical input and music experience descriptors |
US12039959B2 (en) | 2015-09-29 | 2024-07-16 | Shutterstock, Inc. | Automated music composition and generation system employing virtual musical instrument libraries for producing notes contained in the digital pieces of automatically composed music |
US11776518B2 (en) | 2015-09-29 | 2023-10-03 | Shutterstock, Inc. | Automated music composition and generation system employing virtual musical instrument libraries for producing notes contained in the digital pieces of automatically composed music |
US11011144B2 (en) | 2015-09-29 | 2021-05-18 | Shutterstock, Inc. | Automated music composition and generation system supporting automated generation of musical kernels for use in replicating future music compositions and production environments |
US11430419B2 (en) | 2015-09-29 | 2022-08-30 | Shutterstock, Inc. | Automatically managing the musical tastes and preferences of a population of users requesting digital pieces of music automatically composed and generated by an automated music composition and generation system |
US11430418B2 (en) | 2015-09-29 | 2022-08-30 | Shutterstock, Inc. | Automatically managing the musical tastes and preferences of system users based on user feedback and autonomous analysis of music automatically composed and generated by an automated music composition and generation system |
US11468871B2 (en) | 2015-09-29 | 2022-10-11 | Shutterstock, Inc. | Automated music composition and generation system employing an instrument selector for automatically selecting virtual instruments from a library of virtual instruments to perform the notes of the composed piece of digital music |
US11651757B2 (en) | 2015-09-29 | 2023-05-16 | Shutterstock, Inc. | Automated music composition and generation system driven by lyrical input |
US11024275B2 (en) | 2019-10-15 | 2021-06-01 | Shutterstock, Inc. | Method of digitally performing a music composition using virtual musical instruments having performance logic executing within a virtual musical instrument (VMI) library management system |
US10964299B1 (en) | 2019-10-15 | 2021-03-30 | Shutterstock, Inc. | Method of and system for automatically generating digital performances of music compositions using notes selected from virtual musical instruments based on the music-theoretic states of the music compositions |
US11037538B2 (en) | 2019-10-15 | 2021-06-15 | Shutterstock, Inc. | Method of and system for automated musical arrangement and musical instrument performance style transformation supported within an automated music performance system |
Also Published As
Publication number | Publication date |
---|---|
GB201408257D0 (en) | 2014-06-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Senac et al. | Music feature maps with convolutional neural networks for music genre classification | |
Lu et al. | A Newapproach To Query By Humming In Music Retrieval. | |
Davies et al. | AutoMashUpper: Automatic creation of multi-song music mashups | |
Percival et al. | Streamlined tempo estimation based on autocorrelation and cross-correlation with pulses | |
Salamon et al. | Melody extraction from polyphonic music signals: Approaches, applications, and challenges | |
Casey et al. | Content-based music information retrieval: Current directions and future challenges | |
Pohle et al. | Evaluation of frequently used audio features for classification of music into perceptual categories | |
Foote et al. | Audio Retrieval by Rhythmic Similarity. | |
Goto | SmartMusicKIOSK: Music listening station with chorus-search function | |
CN107203571B (en) | Song lyric information processing method and device | |
Mauch et al. | Timbre and Melody Features for the Recognition of Vocal Activity and Instrumental Solos in Polyphonic Music. | |
Marolt | A mid-level representation for melody-based retrieval in audio collections | |
WO2015170126A1 (en) | Methods, systems and computer program products for identifying commonalities of rhythm between disparate musical tracks and using that information to make music recommendations | |
EP3796306B1 (en) | Audio stem identification systems and methods | |
Dittmar et al. | Audio forensics meets music information retrieval—a toolbox for inspection of music plagiarism | |
Rocha et al. | Segmentation and timbre-and rhythm-similarity in Electronic Dance Music | |
Schuller et al. | Determination of nonprototypical valence and arousal in popular music: features and performances | |
US9053695B2 (en) | Identifying musical elements with similar rhythms | |
EP3796305B1 (en) | Audio stem identification systems and methods | |
Paiva et al. | On the Detection of Melody Notes in Polyphonic Audio. | |
Grosche | Signal processing methods for beat tracking, music segmentation, and audio retrieval | |
Zhu et al. | Perceptual visualization of a music collection | |
Gainza et al. | Tempo detection using a hybrid multiband approach | |
Lidy | Evaluation of new audio features and their utilization in novel music retrieval applications | |
Wu et al. | Automatic emotion classification of musical segments |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15721816 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15721816 Country of ref document: EP Kind code of ref document: A1 |