There are two ways of accessing the musical content of a work.
Using various audio-based data extraction tools such as Aubio (http://aubio.org/), Yaafe (https://github.com/Yaafe/Yaafe) and Vamp plugins (www.vamp-plugins.org).
Once the data has been extracted, a whole world of brand new analytical possibilities opens up for musicologists, ethnomusicologists and researchers interested in the phenomenon of music. They can analyse musical data to find relationships in the musical structure, such as the most frequent chord resolutions in a particular style or how a series is developed in a twelve-tone work, or look at the differences between hundreds of performances of the same piece of music. One of the commonest techniques in this kind of analysis is to convert the data obtained into visualisations, such as those provided, for example, by the Sonic Visualiser Free Software, whose possibilities go beyond conventional sound recordings: www.sonicvisualiser.org
The Digital Music Lab project (http://dml.city.ac.uk/) contains several examples of musicological research using Big Data, such as its DML Vis interface which is available on-line (http://dml.city.ac.uk/vis/). This application enables users to explore, analyse and compare sound recordings in three major collections (the British Library Sound Archives, CHARM and I like Music) by viewing the duration of movements, the notes which predominate in terms of length, geographical provenance, chords, etc.
New tools have emerged in recent years which allow musical notation to be codified from its digitised image. Optical Music Recognition (OMR) allows large quantities of musical information to be accessed and processed in digital libraries.
Until new optical recognition technologies appear, the key to reducing conversion errors would appear to involve subsequent processing of the codified data.
As a solution to correct the results obtained after performing OMR, the researchers and research groups we mention below propose processing based on the following strategies:
In addition, with manuscript scores (which we assessed with particular care in the tests) the situation is made even more difficult, if that is possible, by the fact that the musical notation itself is not standardised. This problem is in fact common in the repertoire we used as a case study in our library for this project – vocal music for the stage. For example, if one of the pages does not keep the key and time signatures, the musical writing will be so diverse and so specific that it may produce a conversion which is unusable in practical terms. At present, the conversion represented by a digitised manuscript score is completely ambiguous to the OMR software and requires post-conversion manual correction to obtain an accurate representation of the musical notation.