music21.search.base

base classes for searching scores.

See User’s Guide, Chapter 43: Searching in and Among Scores for details.

SearchMatch

class music21.search.base.SearchMatch(elStart, els, index, iterator)

A lightweight object representing the match (if any) for a search. Derived from namedtuple

SearchMatch bases

SearchMatch instance variables

SearchMatch.elStart

The first element that matches the list.

SearchMatch.els

A tuple of all the matching elements.

SearchMatch.index

The index in the iterator at which the first element can be found

SearchMatch.iterator

The iterator which produced these elements.

StreamSearcher

class music21.search.base.StreamSearcher(streamSearch: Stream, searchList: list[music21.base.Music21Object])

An object that can search through streams for a set of elements or notes or something of that sort.

Create a basic Stream first:

>>> thisStream = converter.parse('tinynotation: 3/4 c4. d8 e4 g4. a8 f4. c4. d4')
>>> thisStream.show('text')
{0.0} <music21.stream.Measure 1 offset=0.0>
    {0.0} <music21.clef.TrebleClef>
    {0.0} <music21.meter.TimeSignature 3/4>
    {0.0} <music21.note.Note C>
    {1.5} <music21.note.Note D>
    {2.0} <music21.note.Note E>
{3.0} <music21.stream.Measure 2 offset=3.0>
    {0.0} <music21.note.Note G>
    {1.5} <music21.note.Note A>
    {2.0} <music21.note.Note F>
{6.0} <music21.stream.Measure 3 offset=6.0>
    {0.5} <music21.note.Note C>
    {2.0} <music21.note.Note D>
    {3.0} <music21.bar.Barline type=final>

Let’s create something to search for:

>>> c = note.Note('C', quarterLength=1.5)
>>> d = note.Note('D', quarterLength=0.5)
>>> searchList = [c, d]

Now create a StreamSearcher:

>>> ss = search.StreamSearcher(thisStream, searchList)

searchList could also be a Stream in itself.

Let’s configure it for recursive search and to filter so only notes are there:

>>> ss.recurse = True
>>> ss.filterNotes = True  # or `.filterNotesAndRests`

Alternatively, we could have passed in a StreamIterator instead of thisStream.

Now let’s configure the algorithms:

>>> ss.algorithms
[<function StreamSearcher.wildcardAlgorithm at 0x111b6e340>]

Wildcard search is a default algorithm that lets you use wildcards. I suggest you leave it in place and add to the algorithms list. We can add the rhythmAlgorithm to it:

>>> ss.algorithms.append(search.StreamSearcher.rhythmAlgorithm)
>>> ss.algorithms
[<function StreamSearcher.wildcardAlgorithm at 0x111b6e340>,
 <function StreamSearcher.rhythmAlgorithm at 0x11200000>]

Now run it:

>>> results = ss.run()
>>> results
[SearchMatch(elStart=<music21.note.Note C>, els=len(2), index=0, iterator=[...]),
 SearchMatch(elStart=<music21.note.Note G>, els=len(2), index=3, iterator=[...])]
>>> results[0].elStart.measureNumber
1
>>> results[1].elStart.measureNumber
2

Wildcards can be useful:

>>> searchStream2 = stream.Stream()
>>> searchStream2.append(note.Note(quarterLength=0.5))
>>> searchStream2.append(search.Wildcard())
>>> searchStream2.append(note.Note(quarterLength=1.5))
>>> ss.searchList = searchStream2
>>> results = ss.run()
>>> results
[SearchMatch(elStart=<music21.note.Note D>, els=len(3), index=1, iterator=[...]),
 SearchMatch(elStart=<music21.note.Note A>, els=len(3), index=4, iterator=[...])]
>>> results[0].els
(<music21.note.Note D>, <music21.note.Note E>, <music21.note.Note G>)
>>> [n.duration.quarterLength for n in results[0].els]
[0.5, 1.0, 1.5]

StreamSearcher methods

StreamSearcher.noteNameAlgorithm(streamEl: Music21Object, searchEl: Music21Object)
StreamSearcher.rhythmAlgorithm(streamEl: Music21Object, searchEl: Music21Object)
StreamSearcher.run() list[music21.search.base.SearchMatch]
StreamSearcher.wildcardAlgorithm(streamEl: Music21Object, searchEl: Music21Object)

An algorithm that supports Wildcards – added by default to the search function.

Wildcard

class music21.search.base.Wildcard(**keywords)

An object that may have some properties defined, but others not that matches a single object in a music21 stream. Equivalent to the regular expression “.”

>>> wc1 = search.Wildcard()
>>> wc1.pitch = pitch.Pitch('C')
>>> st1 = stream.Stream()
>>> st1.append(note.Note('D', type='half'))
>>> st1.append(wc1)

Wildcard bases

Wildcard read-only properties

Read-only properties inherited from Music21Object:

Read-only properties inherited from ProtoM21Object:

Wildcard read/write properties

Read/write properties inherited from Music21Object:

Wildcard methods

Methods inherited from Music21Object:

Methods inherited from ProtoM21Object:

Wildcard instance variables

Instance variables inherited from Music21Object:

WildcardDuration

class music21.search.base.WildcardDuration(typeOrDuration: str | OffsetQLIn | DurationTuple | None = None, /, *, type: str | None = None, dots: int | None = 0, quarterLength: OffsetQLIn | None = None, durationTuple: DurationTuple | None = None, components: Iterable[DurationTuple] | None = None, client: base.Music21Object | None = None, **keywords)

a wildcard duration (it might define a duration in itself, but the methods here will see that it is a wildcard of some sort)

No difference from any other duration.

WildcardDuration bases

WildcardDuration read-only properties

Read-only properties inherited from Duration:

Read-only properties inherited from ProtoM21Object:

WildcardDuration read/write properties

Read/write properties inherited from Duration:

WildcardDuration methods

Methods inherited from Duration:

Methods inherited from ProtoM21Object:

WildcardDuration instance variables

Instance variables inherited from Duration:

Functions

music21.search.base.approximateNoteSearch(thisStream, otherStreams)

searches the list of otherStreams and returns an ordered list of matches (each stream will have a new property of matchProbability to show how well it matches)

>>> s = converter.parse("tinynotation: 4/4 c4 d8 e16 FF a'4 b-")
>>> o1 = converter.parse("tinynotation: 4/4 c4 d8 e GG a' b-4")
>>> o1.id = 'o1'
>>> o2 = converter.parse("tinynotation: 4/4 d#2 f A a' G b")
>>> o2.id = 'o2'
>>> o3 = converter.parse("tinynotation: 4/4 c8 d16 e32 FF32 a'8 b-8")
>>> o3.id = 'o3'
>>> l = search.approximateNoteSearch(s, [o1, o2, o3])
>>> for i in l:
...    print('%s %r' % (i.id, i.matchProbability))
o1 0.666666...
o3 0.333333...
o2 0.083333...
music21.search.base.approximateNoteSearchNoRhythm(thisStream, otherStreams)

searches the list of otherStreams and returns an ordered list of matches (each stream will have a new property of matchProbability to show how well it matches)

>>> s = converter.parse("tinynotation: 4/4 c4 d8 e16 FF a'4 b-")
>>> o1 = converter.parse("tinynotation: 4/4 c4 d8 e GG a' b-4")
>>> o1.id = 'o1'
>>> o2 = converter.parse("tinynotation: 4/4 d#2 f A a' G b")
>>> o2.id = 'o2'
>>> o3 = converter.parse("tinynotation: 4/4 c4 d e GG CCC r")
>>> o3.id = 'o3'
>>> l = search.approximateNoteSearchNoRhythm(s, [o1, o2, o3])
>>> for i in l:
...    print('%s %r' % (i.id, i.matchProbability))
o1 0.83333333...
o3 0.5
o2 0.1666666...
music21.search.base.approximateNoteSearchOnlyRhythm(thisStream, otherStreams)

searches the list of otherStreams and returns an ordered list of matches (each stream will have a new property of matchProbability to show how well it matches)

>>> s = converter.parse("tinynotation: 4/4 c4 d8 e16 FF a'4 b-")
>>> o1 = converter.parse("tinynotation: 4/4 c4 d8 e GG a' b-4")
>>> o1.id = 'o1'
>>> o2 = converter.parse("tinynotation: 4/4 d#2 f A a' G b")
>>> o2.id = 'o2'
>>> o3 = converter.parse("tinynotation: 4/4 c4 d e GG CCC r")
>>> o3.id = 'o3'
>>> l = search.approximateNoteSearchOnlyRhythm(s, [o1, o2, o3])
>>> for i in l:
...    print('%s %r' % (i.id, i.matchProbability))
o1 0.5
o3 0.33...
o2 0.0
music21.search.base.approximateNoteSearchWeighted(thisStream, otherStreams)

searches the list of otherStreams and returns an ordered list of matches (each stream will have a new property of matchProbability to show how well it matches)

>>> s = converter.parse("tinynotation: 4/4 c4 d8 e16 FF a'4 b-")
>>> o1 = converter.parse("tinynotation: 4/4 c4 d8 e GG2 a' b-4")
>>> o1.id = 'o1'
>>> o2 = converter.parse("tinynotation: 4/4 AAA4 AAA8 AAA16 AAA16 AAA4 AAA4")
>>> o2.id = 'o2'
>>> o3 = converter.parse("tinynotation: 4/4 c8 d16 e32 FF32 a'8 b-8")
>>> o3.id = 'o3'
>>> o4 = converter.parse("tinynotation: 4/4 c1 d1 e1 FF1 a'1 b-1")
>>> o4.id = 'o4'
>>> l = search.approximateNoteSearchWeighted(s, [o1, o2, o3, o4])
>>> for i in l:
...    print('%s %r' % (i.id, i.matchProbability))
o3 0.83333...
o1 0.75
o4 0.75
o2 0.25
music21.search.base.mostCommonMeasureRhythms(streamIn, transposeDiatonic=False)

returns a sorted list of dictionaries of the most common rhythms in a stream where each dictionary contains:

number: the number of times a rhythm appears rhythm: the rhythm found (with the pitches of the first instance of the rhythm transposed to C5) measures: a list of measures containing the rhythm rhythmString: a string representation of the rhythm (see translateStreamToStringOnlyRhythm)

>>> bach = corpus.parse('bwv1.6')
>>> sortedRhythms = search.mostCommonMeasureRhythms(bach)
>>> for in_dict in sortedRhythms[0:3]:
...     print(f"no: {in_dict['number']} rhythmString: {in_dict['rhythmString']}")
...     print('bars: %r' % ([(m.number,
...                               str(m.getContextByClass(stream.Part).id))
...                            for m in in_dict['measures']]))
...     in_dict['rhythm'].show('text')
...     print('-----')
no: 34 rhythmString: PPPP
bars: [(1, 'Soprano'), (2, 'Soprano'), (3, 'Soprano'), ..., (1, 'Alto'), ..., (10, 'Bass')]
{0.0} <music21.note.Note C>
{1.0} <music21.note.Note A>
{2.0} <music21.note.Note F>
{3.0} <music21.note.Note C>
-----
no: 7 rhythmString: ZZ
bars: [(13, 'Soprano'), (14, 'Soprano'), ..., (14, 'Bass')]
{0.0} <music21.note.Note C>
{2.0} <music21.note.Note A>
-----
no: 6 rhythmString: ZPP
bars: [(6, 'Soprano'), (12, 'Soprano'), ..., (18, 'Tenor'), ... (12, 'Bass')]
{0.0} <music21.note.Note C>
{2.0} <music21.note.Note B->
{3.0} <music21.note.Note B->
-----
  • Changed in v7: bars are ordered first by number, then by part.

music21.search.base.noteNameRhythmicSearch(thisStreamOrIterator, searchList)
>>> thisStream = converter.parse('tinynotation: 3/4 c4 d8 e c d e f c D E c c4 d# e')
>>> searchList = [note.Note('C'), note.Note('D'), note.Note('E')]
>>> for n in searchList:
...     n.duration.type = 'eighth'
>>> thisStreamIter = thisStream.recurse().notes
>>> search.noteNameRhythmicSearch(thisStreamIter, searchList)
[3, 7]
>>> searchList[0].duration = search.WildcardDuration()
>>> search.noteNameRhythmicSearch(thisStreamIter, searchList)
[0, 3, 7]
music21.search.base.noteNameSearch(thisStreamOrIterator, searchList)
>>> thisStream = converter.parse('tinynotation: 3/4 c4 d8 e c d e f c D E c c4 d# e')
>>> searchList = [note.Note('C'), note.Note('D'), note.Note('E')]
>>> thisStreamIter = thisStream.recurse().notes
>>> search.noteNameSearch(thisStreamIter, searchList)
[0, 3, 7]
>>> searchList2 = [note.Note('C'), search.Wildcard(), note.Note('E')]
>>> search.noteNameSearch(thisStreamIter, searchList2)
[0, 3, 7, 11]
music21.search.base.rhythmicSearch(thisStreamOrIterator, searchList)

Takes two streams – the first is the stream to be searched and the second is a stream of elements whose rhythms must match the first. Returns a list of indices which begin a successful search.

searches are made based on quarterLength. thus a dotted sixteenth-note and a quadruplet (4:3) eighth will match each other.

Example 1: First we will set up a simple stream for searching:

>>> thisStream = converter.parse('tinynotation: 3/4 c4. d8 e4 g4. a8 f4. c4. r4')
>>> thisStream.show('text')
{0.0} <music21.stream.Measure 1 offset=0.0>
    {0.0} <music21.clef.TrebleClef>
    {0.0} <music21.meter.TimeSignature 3/4>
    {0.0} <music21.note.Note C>
    {1.5} <music21.note.Note D>
    {2.0} <music21.note.Note E>
{3.0} <music21.stream.Measure 2 offset=3.0>
    {0.0} <music21.note.Note G>
    {1.5} <music21.note.Note A>
    {2.0} <music21.note.Note F>
{6.0} <music21.stream.Measure 3 offset=6.0>
    {0.5} <music21.note.Note C>
    {2.0} <music21.note.Rest quarter>
    {3.0} <music21.bar.Barline type=final>

Now we will search for all dotted-quarter/eighth elements in the Stream:

>>> thisStreamIter = thisStream.recurse().notes
>>> searchStream1 = stream.Stream()
>>> searchStream1.append(note.Note(quarterLength=1.5))
>>> searchStream1.append(note.Note(quarterLength=0.5))
>>> l = search.rhythmicSearch(thisStreamIter, searchStream1)
>>> l
[0, 3]
>>> stream.Stream(thisStreamIter[3:5]).show('text')
{0.0} <music21.note.Note G>
{1.5} <music21.note.Note A>

Slightly more advanced search: we will look for any instances of eighth, followed by a note (or other element) of any length, followed by a dotted quarter note. Again, we will find two instances; this time we will tag them both with a TextExpression of “*” and then show the original stream:

>>> searchStream2 = stream.Stream()
>>> searchStream2.append(note.Note(quarterLength=0.5))
>>> searchStream2.append(search.Wildcard())
>>> searchStream2.append(note.Note(quarterLength=1.5))
>>> l = search.rhythmicSearch(thisStreamIter, searchStream2)
>>> l
[1, 4]
>>> for found in l:
...     thisStreamIter[found].lyric = '*'
>>> thisStream.show()
../_images/searchRhythmic1.png

Now we can test the search on a real dataset and show the types of preparation that are needed to make it most likely a success. We will look through the first movement of Corelli Trio Sonata op. 3 no. 1 (F major) looking to see how much more common the first search term (dotted-quarter, eighth) is than the second (eighth, anything, dotted-quarter). In fact, my hypothesis was wrong, and the second term is actually more common than the first! (n.b. rests are being counted here as well as notes)

>>> grave = corpus.parse('corelli/opus3no1/1grave')
>>> term1results = []
>>> term2results = []
>>> for p in grave.parts:
...    pf = p.flatten().stripTies().notesAndRests  # consider tied notes as one long note
...    temp1 = search.rhythmicSearch(pf, searchStream1)
...    temp2 = search.rhythmicSearch(pf, searchStream2)
...    for found in temp1:
...        term1results.append(found)
...    for found in temp2:
...        term2results.append(found)
>>> term1results
[0, 7, 13, 21, 42, 57, 64, 66, 0, 5, 7, 19, 21, 40, 46, 63, 0, 8, 31, 61, 69, 71, 73, 97]
>>> term2results
[5, 29, 95]
>>> float(len(term1results)) / len(term2results)
8.0
music21.search.base.streamSearchBase(thisStreamOrIterator, searchList, algorithm=None)

A basic search function that is used by other search mechanisms, which takes in a stream or StreamIterator and a searchList or stream and an algorithm to run on each pair of elements to determine if they match.

music21.search.base.translateDiatonicStreamToString(inputStreamOrIterator, returnMeasures=False)

Translates a Stream or StreamIterator of Notes and Rests only into a string, encoding only the .step (no accidental or octave) and whether the note is slower, faster, or the same speed as the previous note.

Skips all but the first note of tie. Skips multiple rests in a row

Each note gets one byte:

A-G = note of same length as previous H-N = note of longer length than previous O-U = note of shorter length than previous Z = rest

>>> s = converter.parse("tinynotation: 3/4 c4 d8~ d16 r16 FF8 F#8 a'8 b-2.")
>>> streamString = search.translateDiatonicStreamToString(s.recurse().notesAndRests)
>>> print(streamString)
CRZFFAI
>>> len(streamString)
7

If returnMeasures is True, returns an array of measureNumbers where each entry represents the measure number of the measure of the object at that character position :

>>> streamString2, measures = search.translateDiatonicStreamToString(s.recurse().notesAndRests,
...                                    returnMeasures=True)
>>> streamString == streamString2
True
>>> measures
[1, 1, 1, 1, 1, 2, 2]
music21.search.base.translateDurationToBytes(n: GeneralNote)

takes a note.Note object and translates it to a two-byte representation

currently returns the chr() for the note’s midi number. or chr(127) for rests followed by the log of the quarter length (fitted to 1-127, see formula below)

>>> n = note.Note('C4')
>>> n.duration.quarterLength = 3  # dotted half
>>> trans = search.translateDurationToBytes(n)
>>> trans
'_'
>>> (2 ** (ord(trans[0]) / 10)) / 256  # approximately 3
2.828...
music21.search.base.translateIntervalsAndSpeed(inputStream, returnMeasures=False)

Translates a Stream (not StreamIterator) of Notes and Rests only into a string, encoding only the chromatic distance from the last note and whether the note is slower, faster, or the same speed as the previous note.

Skips all but the first note of tie. Skips multiple rests in a row

Each note gets one byte and encodes up from -13 to 13 (all notes > octave are 13 or -13)

>>> s = converter.parse("tinynotation: 3/4 c4 d8~ d16 r16 F8 F#8 F8 a'8 b-2")
>>> sn = s.flatten().notesAndRests.stream()
>>> streamString = search.translateIntervalsAndSpeed(sn)
>>> print(streamString)
Ib RHJ<9
>>> print([ord(x) for x in streamString])
[73, 98, 32, 82, 72, 74, 60, 57]
>>> len(streamString)
8

If returnMeasures is True, returns a triplet of whether the last note was a rest, whether the last note was tied, what the last quarterLength was, and what the last pitches’ midi number was

which can be fed back into this algorithm:

>>> streamString2, measures = search.translateIntervalsAndSpeed(sn, returnMeasures=True)
>>> streamString == streamString2
True
>>> measures
[1, 1, 1, 1, 1, 2, 2, 2]
music21.search.base.translateNoteTieToByte(n: GeneralNote)

takes a note.Note object and returns a one-byte representation of its tie status. ‘s’ if start tie, ‘e’ if stop tie, ‘c’ if continue tie, and ‘’ if no tie

>>> n = note.Note('E')
>>> search.translateNoteTieToByte(n)
''
>>> n.tie = tie.Tie('start')
>>> search.translateNoteTieToByte(n)
's'
>>> n.tie.type = 'continue'
>>> search.translateNoteTieToByte(n)
'c'
>>> n.tie.type = 'stop'
>>> search.translateNoteTieToByte(n)
'e'
music21.search.base.translateNoteToByte(n: GeneralNote)

takes a note.Note object and translates it to a single byte representation.

currently returns the chr() for the note’s midi number. or chr(127) for rests and unpitched.

>>> n = note.Note('C4')
>>> b = search.translateNoteToByte(n)
>>> b
'<'
>>> ord(b)
60
>>> ord(b) == n.pitch.midi
True

Chords are currently just searched on the first Note (or treated as a Rest if None)

music21.search.base.translateNoteWithDurationToBytes(n: GeneralNote, includeTieByte=True)

takes a note.Note object and translates it to a three-byte representation.

currently returns the chr() for the note’s midi number. or chr(127) for rests followed by the log of the quarter length (fitted to 1-127, see translateDurationToBytes()) followed by ‘s’, ‘c’, or ‘e’ if includeTieByte is True and there is a tie.

>>> n = note.Note('C4')
>>> n.duration.quarterLength = 3  # dotted half
>>> trans = search.translateNoteWithDurationToBytes(n)
>>> trans
'<_'
>>> (2**(ord(trans[1])/10.0))/256  # approximately 3
2.828...
>>> n.tie = tie.Tie('stop')
>>> trans = search.translateNoteWithDurationToBytes(n)
>>> trans
'<_e'
>>> trans = search.translateNoteWithDurationToBytes(n, includeTieByte=False)
>>> trans
'<_'
music21.search.base.translateStreamToString(inputStreamOrIterator, returnMeasures=False)

takes a stream (or streamIterator) of notesAndRests only and returns a string for searching on.

>>> s = converter.parse("tinynotation: 3/4 c4 d8 r16 FF8. a'8 b-2.")
>>> sn = s.flatten().notesAndRests
>>> streamString = search.translateStreamToString(sn)
>>> print(streamString)
<P>F<)KQFF_
>>> len(streamString)
12

Chords give the pitch only of the first note and Unpitched objects are treated as rests:

>>> s = stream.Stream([note.Note('C4'), note.Rest(),
...                    chord.Chord(['C4', 'E4']), note.Unpitched()])
>>> streamString = search.translateStreamToString(s)
>>> list(streamString.encode('utf-8'))
[60, 80, 127, 80, 60, 80, 127, 80]
music21.search.base.translateStreamToStringNoRhythm(inputStream, returnMeasures=False)

takes a stream or streamIterator of notesAndRests only and returns a string for searching on, using translateNoteToByte.

>>> s = converter.parse("tinynotation: 4/4 c4 d e FF a'2 b-2")
>>> sn = s.flatten().notesAndRests
>>> search.translateStreamToStringNoRhythm(sn)
'<>@)QF'

With returnMeasures, will return a tuple of bytes and a list of measure numbers:

>>> search.translateStreamToStringNoRhythm(sn, returnMeasures=True)
('<>@)QF', [1, 1, 1, 1, 2, 2])
music21.search.base.translateStreamToStringOnlyRhythm(inputStream, returnMeasures=False)

takes a stream or streamIterator of notesAndRests only and returns a string for searching on.

>>> s = converter.parse("tinynotation: 3/4 c4 d8 e16 FF8. a'8 b-2.")
>>> sn = s.flatten().notesAndRests
>>> streamString = search.translateStreamToStringOnlyRhythm(sn)
>>> print(streamString)
PF<KF_
>>> len(streamString)
6