GSoC 2017: Directly Access the MusicBrainz Database in CritiqueBrainz

Tags: #<Tag:0x00007fe3d42bc018> #<Tag:0x00007fe3d42abd80>


Personal Information:

Project Overview:

Presently, CritiqueBrainz uses python-musicbrainzngs to show search results and fetch info of selected entities. python-musicbrainzngs then uses the XML Web Service which returns the requested results. This is not very slow but some pages on CritiqueBrainz require a lot of MusicBrainz data, which takes very long time to retrieve. For example, as suggested in CB-162, if the cache is empty and each user requests separate pages of the review browsing section, then there are 330 X 27 requests to the web service. Directly accessing the MusicBrainz database would mean one query(directly to the database) per page(by getting multiple entities’ data in a batch using raw SQL statements) only if the cache is completely empty. Thus, directly accessing the MB database would result in significant increase in speed with which CB loads.

Migrating off python-musicbrainzngs and directly accessing MB Database:

Migrate off ngs

What I propose to do can be described in the following subtasks:

  1. Setting up the development environment for accessing MB Database: MB database would be made available in the development environment by setting up a separate docker container. This can easily be achieved by using the musicbrainz-test-database docker image and then importing the MusicBrainz data dumps. Using this docker image ensures that future schema changes in MusicBrainz database don’t break the build environment for CB development. Accessing the MB Database would be made an option in config files as some users might not be willing to download the database dumps.

  2. Directly use the MusicBrainz Database: CritiqueBrainz presently allows reviewing release groups, places, and events. Directly accessing the MB database would require writing SQL statements for accessing the info regarding different entities.

  3. Improving User Experience: Issues like CB-80, CB-84, CB-93, CB-140 can be resolved by using the MB Database directly. Consider release groups, data fetched via the WS was difficult to separately sort by release year. Using MB database directly would facilitate doing such things and improving the user experience. Other music review sites like rateyourmusic have artists albums sorted by release dates (see example). CritiqueBrainz presently does not do so (see example). I will also work on issues CB-36 and CB-198 for enhancing ways in which reviews are shown.

Implementation Details:


Using the MB Database would be made a configurable option for development. This can be implemented using separate docker-compose files for the build. For adding the MusicBrainz database docker container, this existing docker image would be used.

The MusicBrainz database will reside in a separate docker container in the development environment. I would then update the CritiqueBrainz build documentation.
The existing docker-compose files would then include making another docker container using the musicbrainz-test-database image. Dumps can be then downloaded and extracted. The database can be then populated using the COPY command that PostgreSQL provides. There are two options for achieving this: a bash script to run in the container or adding function in for adding data dumps to the MB database. Any of the two options can be used. I have started some work on the MB database access part. I would also write functions for validating dumps using md5sums, checking for correct SCHEMA SEQUENCE and TIMESTAMP and checking for disk space on the host device.

Writing Database Queries:

For getting entity info, we could get most information using joins on many tables(for example, considering release groups, information like artists, artist_ids, primary_type of a release_group, release_group_name, release_group_id can be attained at once using tables artist_credit, artist, artist_credit_name, release group and release_group_primary_type). For information like related URLs and tags of entities, we may need to make separate queries. An example function for querying release group info from release group id is shown below. An important point is that release group information needed on the browse-reviews can be easily batched using SQL operators like IN etc. The retrieved data will be cached as it is presently done.

The frontend/external/musicbrainz/ will contain the following files.

  • This file will contain functions like init_mb_db_engine to use SQLAlchemy for querying mb_database.(see example), This file will also contain wrapper functions to decide which functions to call (direct database access or web service based functions) based on the set config. Other modules can then access these functions to fetch the data.
  • This file will contain the existing search functions which query the WS to get search results.

These files will contain the existing functions which access the WS as well as MB Database access functions:

  • This file will contain the following functions:
    • get_artist_by_id: Get information related to an artist like name, artist area, external links etc.
    • get_artist_urls: Get external links of an artist. Also arranges external links of artists into dictionaries with keys like AllMusic, WikiData, etc.
    • get_artist_relations: Get artist relations like members of a band or band of an individual artist. Arranges relations properly into dictionaries.
  • This file will contain the following functions:
    • get_event_by_id: Get information related to an event like event name, artists performed, place etc.
    • get_artist_rels: Get artists that have performed at an event. Also arranges artists into dictionaries with keys like Main Performers, Supporting Acts, etc.
    • get_place_rels: Get place info of where the event was held.
  • Functions like:
    • get_place_by_id: Get information related to a place like location, coordinates, type of the place.
    • get_artist_rels: Get information related artists performed at those places.
      and other similar functions.
  • For fetching information related to relase_groups and releases:
    • get_multiple_release_groups_by_ids: Get information regarding a bunch of release groups. The information returned would include the artist MBID, artist name, release group MBID, release group type and release group name.
    • get_release_group_rels: Get external links of a release group. Also arrange links in dictionaries.
    • browse_release_groups: browse release groups of an artist. The data here would also contain primary types and then would be arranged into a dictionary.
    • browse_all_releases: Get all the releases of a certain artist or a release group.
    • get_release_by_id: Get info about a release like recordings, recordings length etc.
    • get_release_group_by_id: Get all info about release_group. This includes getting external links, tags etc.
  • This file will have helper functions similar to Search Server. These functions can be used to fetch information which is common between entities. (eg. tags)
  • tests/: This directory will contain files,,, for testing functions in above files.

Fetching information of entities and Improving User experience:

After setting up the docker containers and getting things started, I would help setting up access of MB Database on the production server. Any of the running instances of MB Server can be connected to CB. A Mirror server will be the best to connect CB to.

I will write functions to access entity information from the MusicBrainz Database. An example function to fetch information regarding a release_group can be found here. l_*_* tables are used in MB Database to link corresponding rows in entity tables. Also, in MB database, links suggest the kind of relation between the entities. For eg, there are artists which can be the guest or main performers at an event.

I also plan to have a bunch of helper functions that would help to get related information of entities. For example, multiple entities can have tags. Thus, a helper function would be good to have which would accept the entity and relation table name and return tags related to the entity. See example below.

After using raw SQL to access entity info, small tweaks in DB access functions like order by timestamp (For eg. ordering release groups of an artist by release year using table release_group_meta) can be done to solve issues like CB-80. Similar tweaks can be done to show releases with bonus tracks. I plan to solve these and some other issues alongside while writing database access functions to improve the user experience. Tests will be written for all functions also.


A broad timeline of my work during summer can be summarized as below:

  • Community Bonding (May 5 - May 29): Will start setting/completing up the docker containers for the development environment of CB. Also, discuss design decisions with the mentor.

  • Phase I (May 30 - June 20): Complete setting up MB Database for development environment using docker containers. Adding a function to or writing a script for importing mbdumps will also be part of this task. Also, update the CritiqueBrainz build documentation. Also, help setting up changes in the production server. In this phase, I also aim to complete database access functions for places.

  • Phase II (June 21 - July 20): I plan to complete database access functions for release_group and artists. Also, I will modify entity info functions to improve user experience and thus close the associated issues.

  • Phase III (July 21 - August 29): I will complete all database access functions in this phase. I will work on improving user experience and issues for displaying entities only with reviews. Also, spend time fixing bugs.

  • After GSoC: Continue working on CritiqueBrainz and maybe other MusicBrainz Projects. Implementing the rating system in CritiqueBrainz would also be an interesting work to do.

A detailed timeline of my work is as follows:

  • Week 1: Begin with the docker setup of MusicBrainz database for CritiqueBrainz development environment. Make building MB database docker a separate option by using a seperate docker-compose file. Also, alter config files for the option to use the MusicBrainz database. Start with functions to check the md5sums of dumps, SCHEMA SEQUENCE checks` and other tasks.
  • Week 2: Complete with the docker setups. Update build documentation for CritiqueBrainz. Complete functions or script for importing mb_dumps to the docker container.
  • Week 3: Help setting up changes in the production server. Also, begin with DB access functions for places.
  • Week 4: Complete DB access functions for places. Also, begin with the DB access functions for artists.
  • Week 5: Complete DB access functions for artists. Write tests simultaneously.
  • Week 6: Write functions for accessing information regarding release groups.
  • Week 7: Complete writing functions for release and release group information. Keep tweaking functions for improving results displayed.
  • Week 8: Complete any pending stuff. If not, start with functions for events information. Complete functions for fetching event info (and tests simultaneously). Improving UX stuff by ordering releases and release group information according to desired sort(release dates, for instance).
  • Week 9: Complete previous week’s tasks and start working on filters to show review only entities.
  • Week 10: Complete any pending stuff.
  • Week 11: Work more on user experience issues.
  • Week 12: Complete previous week’s tasks. Solve bugs.
  • Week 13: Complete any pending stuff. Work on final submission.

About me:

I am a Computer Science student at National Institute of Technology, Hamirpur. Here is a link to my contributions to CritiqueBrainz.

Tell us about the computer(s) you have available for working on your SoC project!

I presently have a Lenovo laptop with 2 GB RAM and Intel Pentium P6200 processor running Arch Linux. Also, I use vim for all my development.

When did you first start programming?

I started programming in high school writing small C++ programs. Started programming in Python as a freshman in college.

What type of music do you listen to? (Please list a series of MBIDs as examples.)

I listen mostly to rock, switching to metal and pop occasionally. I love listening to songs by Linkin Park, Foo Fighters, Radiohead, Muse, Coldplay, Metallica, Guns N Roses and Thiry Seconds to Mars. My top listens presently are Sultans of Swing and Everlong.

What aspects of the project you’re applying for (e.g., MusicBrainz, AcousticBrainz, etc.) interest you the most?

I love music and love to read reviews of other people about the music I hear and the music that interests me. Reviews submitted to CritiqueBrainz can be used for various purposes such as ranking albums, sentimental analysis etc. This is why I am interested in CB.

Have you ever used MusicBrainz to tag your files?

Yes, I’ve been using Picard to tag my music collection for a few years.

Have you contributed to other Open Source projects? If so, which projects and can we see some of your code?

I have contributed to GNOME Music and Three patches to GNOME can be found here, here and here and contributions to can be found here. I have also made a small contribution to MusicBrainz server and a small build fix in ListenBrainz.

What sorts of programming projects have you done on your own time?

I have made a working auto drive car using the Raspberry Pi (, also made a website in flask called Sentikart for sentimental analysis of user reviews on e-commerce sites. I have also worked on a simple music genre classifier GenreM using Logistic Regression. My other projects include a self-playing hex game bot which uses minimax to decide moves.

How much time do you have available, and how would you plan to use it?

I have holidays during most of the coding period of GSoC and would be able to give 40-50 hrs per week to my project.

Do you plan to have a job or study during the summer in conjunction with Summer of Code?

Not if I get selected for GSoC.

GSoC 2018: More detailed integration of AcousticBrainz with MusicBrainz

This is a draft of my GSoC proposal. Reviews and Suggestions would be appreciated :slight_smile:


My initial reaction is that this sort of direct database use would immediately make any MB schema changes much harder.
If the current MB WS API has issues for CB’s use cases, it makes more sense to me to add some additional endpoint(s) there (or parameters on existing ones), to make those use cases work better.


I don’t see this being very scalable. Isn’t one of the reasons why direct MB db access is wanted, that we’d be able to do more things/get more (/specific) data? If this is not possible with the WS, then you would have to do one thing in “db mode” and another in “ws mode”. Also, since would be using the “db mode”, “ws mode” wouldn’t really be exposed much and would thus likely bitrot fast and mostly be a maintenance nuisance.

I’d say pick one of the two and go with that. Supporting both is not going to lead to much fun.


This was suggested as an option by Gentlecat here. However, if preferred we can definitely remove WS entirely. :slight_smile:


Not necessarily. This much detail is probably not necessary. You can remove the “Search Server” part. All we need to know in the context of CB is that MusicBrainz has an API, which we can access, and that this API somehow queries the database in the backend.

This shouldn’t matter if MusicBrainz database is easy to set up with some data in it. In fact, this should be an easy process for other development tasks too. Can you explain how empty MusicBrainz database is going to be populated with data?

Do you know about existence of Can we use this for development and testing. I’d prefer if we could reuse a universal image. That way it would be easier to keep up to date whenever relevant parts of the MusicBrainz database schema change.

A thing worth noting is that there should be one interface for modules that need to retrieve data from MusicBrainz. Data can be retrieved either from the web service or the database directly (depending on configuration), but users (other modules) shouldn’t have to worry about that.

Not sure if this would be a good idea. I might prefer to just keep using existing search interface in the web service. @Bitmap, any thoughts about this?

What do you mean here?

I would say that it’s better to do this as soon as you have something working. You probably won’t be able to set up production server yourself, but me or someone else would be able to help with this step.

Perhaps it’s better to leave this out. If you don’t know for sure if you are going to do it, then there’s no need to put it into a proposal. I image a proposal having a clear outline of activities from start to finish of the coding period.


I think changing the web service would be much more difficult and time consuming, considering the resources that we have right now. I don’t see how changes to the MusicBrainz schema would become much harder after connecting CritiqueBrainz directly to it. As long as changes are know enough time in advance and assuming that they are affecting the way CritiqueBrainz retrieved data from the MusicBrainz database, I don’t see this being a problem.

New search server, for example, relies on @ferbncode, you might want to look at that project to see if it would be relevant to what you want to do.


Based on what I understand of the proposal, I don’t see any reason to replace the existing search code. The problem of needing to retrieve too much data doesn’t seem to apply to search pages.

The only reason given is that it “would also enable us to rank searches,” but the XML API already returns results ordered by score, descending. I don’t see any other technical advantages given in the proposal, and I can think of a few disadvantages.


There are a couple problems in the written get_release_group_by_id function here:

  • release_group.type is nullable, so you need a left join.
  • release_group.artist_credit and are not the same thing. An artist credit links multiple artists through the artist_credit_name table. If you want to get all of these in one query, you’ll probably need a GROUP BY.


Thanks for the review :slight_smile:

The setup gets much easier using the musicbrainz-test-database, as you had suggested. Database dumps provided can be used to populate the database.
Specifically, I think this can be achieved using the COPY command in PostgreSQL. After extracting the dumps, the database can be populated using:

psql -d musicbrainz_test -U musicbrainz -a -c "\COPY $tablename FROM '$file'

Also, after further thought, since no write updates are made to the database, do we require a separate test database?. Could the same database be used for testing purposes? Please suggest :slight_smile:

Considering your suggestions, I have updated my proposal. The frontend/external/musicbrainz will have files containing both the WS based as well as the MB database functions. We can then have wrappers (which decide functions called based on the config) in a file, which would be then used by other modules.

I have updated the proposal accordingly.

I wanted to actually state that indexes are present in MB database and we don’t have to worry about creating indexes for fast access. To increase clarity, I have omitted this part.

I have updated the proposal to help set up changes in production after docker setup and omitted the optional ideas part.


Whoops, sorry for the mistake.:frowning: I’ve fixed it. Thanks for pointing it out.