Empowering the Next Generation of Women in Audio

Join Us

Boosting Women’s Voices: Cutting Through The Noise

When it comes to editing voices, it’s a job filled with variety, constantly reacting to what hits the ears. While an initial setup of EQ templates might be a starting point for some, every voice is unique. Women’s voices tend to have wildly different tones and timbres that vary from person to person, and editing seems to be an area that’s often hit-and-miss across music and the spoken word. The NCBI Library of Medicine states that the male speaking voice averages around 60 – 180Hz, while the female voice generally sits around 160 – 300Hz, with roughly an octave’s difference in pitch. Despite this, there seems to be a wild disparity in how women’s voices are treated in general. Perhaps the most common problem can be summarised as cutting too much in the lower areas, and boosting too much in the higher areas when women’s voices are in the mix.

Spoken word

With the podcast industry booming, it’s interesting to observe the difference in the editing of women’s voices compared to men’s. The lack of De-esser treatment, and the copious boosting of high-end frequencies often lead to distraction with every ‘t’ and ‘s’ sound that occurs. Sibilance and harshness can abound, and pull us away from what women are actually saying.

Diagram of the Fletcher-Munson Curve

The Fletcher-Munson Curve measures how our bodies perceive loudness. It is also often referred to as the “equal loudness contour”. Created by Harvey Fletcher and Milden A. Munson in the 1930s, the pair demonstrated how loudness affects the human ear at different frequencies, and where we would perceive (or feel) these pitches and volumes as unpleasant. The most sensitive of these frequency areas that offends the ears lies between 3 – 5kHz, which is the danger zone for sibilance.

Business titan Barbara Corcoran is a fantastic speaker and all-around inspirational career woman. Her voice naturally leans to the high end in pitch and tone and has a propensity for sibilance. When I’d previously watched her on the television show Shark Tank, it was clear that this was her vocal sound, yet when I recently listened to her as a guest on a podcast, I was saddened to hear the edit of Barbara’s voice was jarring in the high-end, and desperately needed a De-esser. I was curious to see how closely my perception of the sound was aligned with what was measurably coming out, so I decided to analyse the podcast in contrast with another recording. I used a Spectral Analysis tool, capturing a snapshot of a word with an ‘s’ sound to compare the two different recordings as fairly as possible, and listened through the same speaker.

Barbara speaking at a TEDx Talk

 

I first measured Barbara speaking at a TEDx Talk. There was definitely a slight peak in the range of 3-5kHz when measuring Barbara’s talk, however, the peak was only a little above the others, notably its neighbour around 2kHz, and again a little above the 500Hz peak. Audibly, the voice still sounds high and naturally sibilant, however, there is a softness to the ‘s’ sound that does not detract from the talk.

In the bottom graph, the peak is marked around the 3 – 5kHz range and stands alone above the peaks in lower ranges, which demonstrates that this problem area is in fact considerably louder than the other frequencies, and not just perceived to be louder and distracting by the ear.

 

Diagram Barbara Corcoran’s voice in the TEDx Talk (top image) versus as a podcast guest (bottom image). 

Diagram Barbara Corcoran’s voice in the TEDx Talk (top image) versus as a podcast guest (bottom image).

 

Music

In music, the same problems surround women singers. Often, in striving to add ‘air’ or ‘brightness’ or ‘clarity’ to a vocal, women’s voices succumb to the harshness in the 3 – 5kHz range. In boosting above 2kHz a little too liberally, and adding reverb or other effects that can further highlight the high-end, women’s voices can end up sounding thin, jarring, and full of squeaky ‘s’ sounds. So how do the experts celebrate the richness and full tonal spectrum of strong women’s vocals, and do it so well?

In a 2011 interview talking about the making of Adele’s album 21, producer Paul Epworth and mix engineer Tom Elmhirst gave a run-down of their process. The pair have worked with some formidable women’s voices, from Florence + The Machine and Amy Winehouse to Adele. On the song Rolling In The Deep, Elmhirst used the Waves Q6 EQ on the chorus vocal, pulling out certain frequencies “very, very heavily”:

“I had the Q6 on the chorus vocal, notching out 930, 1634, and 3175 Hz very, very heavily: -18dB, -18dB, and -12.1dB respectively, with very narrow Q. I also had the EQIII on the lead-vocal sub, notching something out again. Something obviously needed to be taken out. The vocal is the most important thing in the track, and taking those frequencies out allowed me to keep it upfront in the mix, particularly in the chorus. Regarding the outboard, I had the Pultec EQ, Urei 1176, and the Tube-Tech CL1B on the lead vocal sub-insert. The Pultec boosted around 100Hz and 12k. It’s colourful, but not drastic. There was not a lot of gain.” 

 

Diagram of Adele Vocal EQ

 

When it came to De-essers, Elmhirst likes to add several for precision – on Rolling In The Deep, Elmhirst explained:

“I did use two Waves De-essers, one taking out at 5449Hz and the other at 11004Hz. Rather than use one to try to cover all the sibilance I used two. I do that quite often.”

While on Someone Like You, he went even further, summarising his EQ and De-esser decisions on the piano-vocal track:

“I had three de-essers on the lead vocal in this case, working at 4185, 7413 and 7712 Hz, and I did some notching on the Waves Q10, taking out 537, 2973, and 10899 Hz, with maximum Q in all cases. The Sonnox Oxford EQ simply takes out everything below 100Hz, and it adds a little around 8k.”

Boosting women’s voices

It’s interesting to compare and contrast the rich tapestry of content that is available to us these days, as well as the amount of guidance that is out there. Considering women’s speaking voices sit around 160 – 300 Hz it’s staggering how many guides and training materials generally recommend using a low pass filter cutting up to 200 Hz – where the voice actually is – and boosting from 4 kHz and up – where madness lies. Every voice needs something different, whether softly spoken, cutting through in an arrangement, or leading a band at a show.

Delving Into De-mix Technology

Since Peter Jackson’s Get Back documentary about The Beatles, the use of De-mix technology has been more prominent in the public realm, and it is a truly intriguing technique. Even the term ‘De-mix’ is a fascinating one, that mentally evokes a challenge similar to ‘un-baking’ a cake.

In fact, the process of De-mixing has been used by Abbey Road Studios for some time, and the technique was developed in partnership with mastermind technical analyst James Clarke, who recalled that the idea first came to him back in 2009. The first project Clarke created was in 2011, with the reimagining of The Beatles Live At The Hollywood Bowl, with many classic Beatles records subsequently following, including A Hard Day’s Night (the movie), 1+, parts from Sgt Pepper’s Lonely Hearts Club Band 50th Anniversary and The White Album. Aside from The Beatles, David Bowie’s Life on Mars, Rush 2112 – The Live Concert, and material from Cliff Richard and The Shadows, as well as Yusuf/Cat Stevens, have been similarly reworked with the De-mix technology.

What is De-mix technology and how does it work?

Abbey Road Studios explains on their website that in its simplest form, the software enhances the original vocals and helps to amplify the bass, which is something that mixes in the late 1960s were often unable to do.

“Using algorithms that are trained on target instruments, De-mix can extract these components to enhance or reduce targeted EQ or isolation. Not only can De-mix be used to adjust the levels of musical elements within a mix, it can also make vocal isolation or removal a reality.

The new process unlocks mono recordings or those where full multi-tracks do not exist, allowing our engineers to adjust the balance and levels of instruments and vocals within a track to rebuild, rebalance and remix the recording. For remastering projects, De-mix allows our engineers to perform targeted EQ balancing. For example, the engineer can adjust and EQ a bass guitar without any impact on the vocals or drums.”

Abbey Road engineer Lewis Jones talked about working on vintage tapes by The Rolling Stones back in 2018, likening the De-mix process to remastering – he drew similarities between taking an initial stereo track, and then making a multi-track of that stereo in order to edit the parts more individually, and enhance them. In the case of these older tracks, however, the source is more often a mono track, which was commonplace in the 1960s.

The comparison to the remastering process makes the technology a little easier to digest. Delving deeper into the science of how De-mix works, Clarke explains:

“The process is that you create spectrogram models of the target instrumentation you’re looking for, so vocals, guitars, bass, drums, stuff like that. And then the software starts to look for patterns within the mixed version that matches those models. It then creates what are called masks, which effectively, think of them like a specific sieve, you just drop the audio through it and the mask catches the bits it wants to keep and lets everything go through. It then does the same for all the other instruments and eventually, it works out that this bit of audio belongs to the drum, or the vocal or bass guitar.”

Clarke also explained that if engineers were having issues while working with the De-mix software, he could tweak the code and the models to assist the process. And looking to the future, Clarke says he is currently moving into a deep learning approach that uses the same concept of generating these masks to un-mix the audio, however, the masks are learnt rather than derived and can be applied to any song. He states that “It’s producing some stunning results at the moment”.

What could be the impact of De-mix technology?

There appears to be only positive potential in the use of De-mix technology, the most notable being the restorative nature of its application – old, forgotten, or bootlegged tracks can benefit hugely from these techniques, and become resurrected to live a second life.

Abbey Road Studios already offer the De-mix service to clients as a remix or remastering option, and the possibilities for the future usage, licensing, or commercialisation of this technique look promising; should Clarke’s deep learning approach continue to create new versions of De-mix, it seems feasible that the technology could one day become widely available to producers and creators. If it can eventually be used as an adaptive preset (as Clarke described in his description of the technology’s deep learning potential), the impact would be huge. Ultimately, taking the innovation and quality of the Abbey Road techniques, and making the software available to use on records everywhere, is a very exciting prospect.

Energy Conservation in Pop Music

I began using the law of conservation of energy to visualize the evolution of Pop music when I was in high school. I was studying higher-level chemistry and music history as part of my International Baccalaureate degree, and while it felt like two unrelated subjects, I was eager to make a connection between them. In general terms, the law of conservation of energy states that energy in a system can’t be created nor destroyed, it just transfers from one form of energy to another. I started visualizing this idea in musical expression while diving into Western European art music and the evolution of Jazz music in America. I noticed how often others recognized these vast musical genres as being more complex than Pop music, primarily because of how they used more intricate harmonies and orchestration. I thought to myself how unfair it was to label the Pop music I was growing up with as simple. I knew it wasn’t any less than the music that came before it, but I couldn’t articulate why. If everything else in the world around me was advancing, then how could music become more basic?

Imagine that Bach’s “Fugue in G minor” and Taylor Swift’s “Out of the Woods” are both their own “isolated systems” with energy, or in this case, musical expression that transfers between different musical elements in each system. In my opinion, both pieces are complex and full of musical expression, and they hold similar amounts of kinetic energy and potential energy. However, the energy is active in different elements of the songs, and for “Out of the Woods,” that element comes from a technological revolution that Bach had no access to. Bach’s fugue holds most of its musical energy in the counterpoint: the melodic composition and modulation drive the expression of the piece. Meanwhile, most of the musical expression in “Out of the Woods” comes from the interesting sonics of the music production, which is true for a lot of Pop music today. Many Pop songs have simplistic chord progressions, which I think is okay because now the energy resides in sound design, music technology, how something is mixed, or how a vocal is processed. I believe that what we’ve experienced as music evolves is a transfer of energy from composition to production because we have the means to do so.

Let’s look at some excerpts of the sheet music from both pieces stated above. Clearly, one melody is more varied and ornamented than the other. Most of Swift’s song is a singular note with little to no melodic contour and a simple I-V-VI-IV chord progression, while Bach’s composition highlights an intricate subject and countermelody with more advanced modulations. Now let’s imagine what the Pro Tools sessions for both songs might look like. Oh right, Bach didn’t have Pro Tools! The earliest known recordings come from the late 19th century, far past his lifetime, so he likely didn’t even consider the kind of microphone he could use or how he could compress the organ with a Teletronix LA-3A or create an entirely new sound on a synthesizer for the fugue. The energy of the piece is most active where Bach’s capabilities and resources exist: his understanding of advanced harmony and his performance technique. Had Taylor Swift been composing at a time when music production wasn’t really a thing, she might have songs with eclectic modulations and contrapuntal bass parts. However, with Jack Antonoff’s exciting vocal arrangement and sound design for electronic drums and synths, there’s already so much energy in the song, that the harmony in this piece doesn’t need to work as hard. Ultimately, I experience both Bach’s fugue and Swift’s single as having the same amount of musical energy, but the energy is utilized in different parts of both systems.

I know this argument all seems convoluted, but this concept has really helped me in my critical listening. When I listen to any recording, I ask myself, “Where has the energy transferred in this piece, where is it being held, and how is it serving the song?” Sometimes the answer is not the same within one genre or even within one artist. If an element in a song feels simple, we can break it down to its core elements to find where the energy is. It can be in the rhythm, the performance, the sampling technique, or the lyricism to name a few. When I write and produce, I approach a new song with this mentality too. Where do I want the energy to be, where can I simplify the song to let that element shine, and how does it work with the narrative of my song?

An Introduction to Classical Music Production

Many classical musicians have been dedicated to their craft since childhood: they’ve spent thousands of hours perfecting their playing technique, studying theory and harmony and history of music, taking lessons with awe-inspiring (and occasionally fear-inducing) professors, and developing a richness of sound that can fluctuate deftly between dramatic passion and subtle nuance, to make even the most hardened of hearts shed a tear of emotion at such sonic beauty! How do we capture in audio the complex compositions of classical music and the natural resonance of these acoustic instruments, and do justice to the sound that classical musicians and singers have worked so hard to create? Goodbye overdubs and hardcore effects processing: classical music recording and production is generally all about finding the most flattering acoustic space to record in, and capturing the musical instrument or voice in a way that best brings out its natural sonic qualities.

Recording session with chamber music ensemble Duo Otero.

Pre-production

One of the most important aspects when planning a classical music recording is finding a space with acoustics that will cradle the music in a glorious bath of reverb – not too much and not too little. When recording many other genres, we’re often striving for a dry-as-a-bone, deadened studio acoustic that will give us the most control over the sound so we can shape it later. Classical music, on the other hand, doesn’t require overdubbing, and so it’s in our best interest to record it in a nice-sounding space. For example, when listening to a live choral performance, isn’t the experience made so much better by those epic cathedral echoes? We also need to find a quiet place without too much external noise – there’s nothing more annoying than having to stop recording five times because five fire trucks have decided to pass by just at that moment! It’s important to do some research on the instruments and the music to be recorded, to be able to prepare an appropriate recording setup. Whether it’s a solo instrumentalist or a full opera production with orchestra will affect our choice and placement of the microphones, and the number of inputs needed.

Recording

Our aim is to capture a performer or ensemble playing together in a great-sounding acoustic – so the workflow is more linear or horizontal than it is vertical. We’re not overdubbing and layering new sounds on top, but we can capture several takes of the same music and then join the best takes together until we have the whole piece, so it sounds like one performance. Because of this way of working, it’s essential that the performer is as well-prepared as they can be, as we can’t make detailed corrections of pitch or timing as we can in other genre recordings (autotune is a no-no!). As we’re recording natural acoustic sounds that can’t be “fixed in the mix” (did I mention no autotune?), it’s important to choose microphones and pre-amps that will do an excellent job of capturing that audio faithfully without colouring the sound too strongly. When placing microphones, we should think about how and where the sound of an instrument is generated, and how it resonates in the acoustic space. A common basic technique is to use a stereo pair of microphones to capture a musician or a whole ensemble within its acoustic, and then to add “spot mics” – microphones placed closer to individual instruments – to capture more details. If there’s the luxury of an abundance of microphones, we might sometimes add an extra pair of microphones even further away from the sound source to capture more of the acoustic space, and then we can blend all of these microphones together, to taste.

Post-production

Mixing classical music usually involves finding a pleasing balance between the recorded channels (for example, the stereo pair, spot mics, and ambient mics), applying suitable panning, noise reduction, and light EQ, and limiting as necessary (perhaps compression for overly-excited percussion or other highly dynamic instruments). If it’s a large ensemble recording, we might use automation to bring up solo parts if they are shy and need a little help, or to highlight interesting musical details and textures. Often using a touch of digital reverb can add a smooth and satisfying sheen. Especially if a perfect-sounding recording space is just not available (it often happens): some epic digital reverb can help to glow up a flat and boring-sounding space.

Aside from live concert recordings, a lot of classical music post-production lies in the editing: often there’ll be several takes of the same material, and the challenge is to select the best performances and stitch it all together in a seamless way so that the transitions can’t be heard – while maintaining the original energy and pacing of the performance, and not going overboard on crazily detailed editing, as that’s kind of cheating (see TwoSetViolin’s hilarious video 1% Violin Skills 99% Editing Skills)! It is an advantage – and probably essential in some situations – to be able to read music scores. It’s really helpful to follow the score as the musicians are playing, to write notes on the best (and worst) takes, to guide them and suggest what they might like to repeat, change or improve, and to make sure that all parts of a piece have been recorded.

In summary

The world of classical music production is an exciting space where audio engineers, producers, and musicians collaborate closely together to immortalise wonderful compositions in audio so that a wider audience can hear and enjoy them. If you’d like to get into classical music production, there’s no better way than to learn by jumping in and practising doing lots of recordings – try different mics and positions in different acoustic spaces, listen to lots of classical music recordings, read up on the different instruments, and use your ears as your most important tool. You’ll soon be Bach for more!

 

 

An Introduction to Sync Licensing

As a musician, you have most likely become aware of the word “sync”. Perhaps you have researched and feel you have a pretty good understanding of the basics. Maybe you’ve even had sync placements. This blog is going to cover the basics for those who are just hearing the term, but more importantly, I want to help you figure out if sync is for you. In my opinion, there are two clear pathways for a musician to take when it comes to seeking out sync licensing opportunities. Hopefully, this will help you determine if one of those paths is right for you.

First, some definitions

Sync is short for “synchronization” or “synchronization licensing” which is referring to the license music creators need to give to folks who want to “synchronize” video of any kind to recorded music.

Music Supervisor is the person who chooses music for every moment of a film or show. Sometimes the composer and the supe are the same person (lower budget films, usually)

A Music Library is like a library of music. People searching for music can search the database, filtered by various features, such as mood, tempo, genre, female vocal, male vocal, instrumental, and so on.

A Sync Agent is a person (can be independent or work for an agency) that is like the go-between for music supes and musicians. They will often take a cut of the sync fee and might also work in a percentage of the master use.

Production Music is the common term used for “background music”, but may have vocals. Music libraries will often compile “albums” of production music by theme, a specific mood or genre, etc. The licensing is already handled with the creator, which makes it much easier for music supes to quickly select a song without having to wait for agreements, approvals, etc.

What are “songs” that “work for sync”?

When music supervisors are looking for music, they are looking for a certain type of energy, a mood, a feeling. Surprisingly, a good sync song may not necessarily be a “hit” song and a “hit” song may not necessarily work for sync. Once in a while, a hit song is also a great sync song, but that is not the norm. Either way, when the perfect song is found for a particular scene or ad, magic can happen.

The best way to really understand what sync is all about is to do a little observation exercise. You are simply going to observe your normal day of Netflix watching or whatever way you watch your shows. Only today, pay attention to the music being played in conjunction with whatever you are watching. Whether it be a movie, a documentary, a reality show, a TV show from the ’80s, pay attention to the music. How many snippets of songs do you hear in each episode? Do any of the songs sound like “radio” songs? How many are instrumentals? Now, what about the ads? I don’t watch regular TV anymore but I do have a few shows that I love to watch on YouTube. So I still see ads quite regularly. How about you? What kind of music are you hearing in the ads?

Every piece of music you heard was composed, written, performed by a person or people. Each piece of music supposedly has a proper license. A cue sheet was also submitted to a songwriting organization so that the songwriters and publishers can be paid a royalty. The value of that piece of music varies from a penny to hundreds of thousands of dollars and everything in between. The amount paid is based on numerous factors; is it background or under dialog, is it playing on the radio or jukebox on screen, is it with vocals, without vocals, how much of the song is played, where in the film, such as opening credits, montage scene, etc., and is it a well-known song or major artist or an indie? Sooooo many factors play into the “value” of that placement. Some songs are paid an upfront sync fee in addition to the songwriting/publishing royalties. Some are not. Some of the songs (especially background, instrumental music) are composed by someone who might work directly for the company creating the show/movie, or the composer may work for the publisher or library that licensed the music. It’s a complex biz.

So, what about the two paths?

I landed my first sync placement back in 2006 ish and it was sort of a fluke. The long story short is that a co-writer/co-producer and I wrote the song specifically for a small, independent film after reading the film synopsis. The song made it into the movie, which aired on ABC Family (and is still streamed regularly on a variety of platforms). Then we shopped it to some music libraries and a music publisher. One of the libraries secured multiple placements for that song, plus several other songs we had already written. In recent years, I’ve tried to dive deeper into creating specifically for sync but seem to have no time for that. I’ve become crazy busy as a full-time music producer for artists. This has helped me clearly see the two paths.

Path one: you are a creator of hundreds of songs, beats, tracks and are pitching almost as much as you are creating. In this path, it is a numbers game. It’s all about quantity. The more “content” you have, the better your chances of getting a sync placement. This scenario is ideal for you if you;

On this path, you can start out by pitching to music libraries but the ultimate goal will be to network to the point where you are receiving briefs from sync agents and production companies directly. This path can take a lot of time before you begin to see the fruits of your labor. Time is needed to make connections, to find good collaborators, and to earn the trust of sync agents and libraries.

Path two: you are an artist who is focusing on building your artistic brand, creating songs that connect with your fan base, creating music that you love and intend on performing. OR you are like me and love producing with and for artists to help them build their career as artists. This path is ideal for you if you:

If this is the path for you, pitching to a sync agent or a manager or producer who has connections to sync agents or music supervisors may be your best bet. If your genre is very current, it may have a short shelf life so get going on that pitching asap. This path requires that you focus on the main goal (building your music business as an artist and/or producer) and perhaps spend a few hours a week on pitching, emails, metadata, and contracts.

If you are on path two, you can try your hand at creating a song or two that are “you” as an artist and could be released on an album or as a single but would also work for sync. There’s nothing wrong with that approach! How do you know if your song would work for sync? Remember the statement above about songs for sync needing to capture a mood, etc? This is very important. What is also important is that there are no lyrics that are about a specific time, location, person, etc. Once in a while, a song with specific lyrics can work perfectly for a scene but it’s better to keep the lyrics “generic” enough to increase your chances for placement. Generic doesn’t mean boring! This is the actual struggle! Writing lyrics that are genuine, interesting, engaging but not specific is actually the hardest part.

Important Companies, Contacts and Resources:

If you are wondering if your songs are “sync ready”, I’m happy to give them a listen and throw my opinion at you. 😉 Send me up to 3 songs and put “Sync Songs?” in the subject line, to becky@voxfoxproductions.com

 

Demystifying Loudness Standards

               
Every sound engineer refers to some kind of meter to aid the judgments we make with our ears. Sometimes it is a meter on tracks in a DAW or that session’s master output meter, other times it is LEDs lighting up our consoles like a Christmas tree, sometimes it is a handheld sound level meter, other times a VU meter, etc. All of those meters measure audio signal using different scales, but they all use the decibel as a unit of measurement. There is also a way to measure the levels of mixes that are designed to represent the human perception of sound: loudness!

Our job as audio engineers and sound designers is to deliver a seamless aural experience. Loudness standards are a set of guides, measured by particular algorithms, to ensure that everyone who is mixing audio is delivering a product that sounds similar in volume across a streaming service, website, and radio or television station. The less work our audiences have to do, the better we have done our jobs. Loudness is one of the many tools that help us ensure that we are delivering the best experience possible.

History           

A big reason we started mixing to loudness standards was to achieve consistent volume, from program to program as well as within shows. Listeners and viewers used to complain to the FCC and BBC TV about jumps in volume between programs, and volume ranges within programs being too wide. Listeners had to perpetually make volume adjustments on their end when their radio or television suddenly got loud, or to hear what was being said if a moment was mixed too quietly compared to the rest of the program.

In 2007, the International Telecommunications Union (ITU) released the ITU-R BS 1770 standard; a set of algorithms to measure audio program loudness and true-peak level. (Chueks Blog.)  Then, the European Broadcast Union (EBU) began to work with the ITU standard. Then EBU modified their standard when they discovered that gaps of silence could bring a loud program down to their specifications. So they released a standard called EBU R-128. Levels below 8 LUFS of the ungated measurement do not count towards the integrated loudness level, which means that the quiet parts can not skew the measurement of the whole program. The ITU standard is still used internationally.

Even after all of this standardization, television viewers were still being blasted by painfully loud commercials.  So, on December 13th, 2012, the FCC passed the Commercial Advertisement Loudness Mitigation Act. From the FCC website: “Specifically, the CALM Act directs the Commission to establish rules that require TV stations, cable operators, satellite TV providers or other multichannel video program distributors to apply the ATSC A/85 Recommended Practice to commercial advertisements they transmit to viewers. The ATSC A/85 RP is a set of methods to measure and control the audio loudness of digital programming, including commercials.  This standard can be used by all broadcast television stations and pay-TV providers.”    And yup, listeners can file complaints to the FCC if a commercial is too loud. The CALM Act just regulates the loudness of commercials.

Non-Eurocentric countries have their own loudness standards, derived from the global ITU R B.S 1770. China’s standard for television broadcast is GY/T 282-2014; Japan’s is ARIB TR-B32; Australia’s and New Zealand’s is OP-29. Many European and South American countries, along with South Africa, use the EBU R-128 standard. There’s a link with a more comprehensive link at the end of this article, in the resources section.

Most clients you will mix for expect you, the sound designer or sound mixer, to abide by any one of these standards, depending on who is distributing it. (Apple, Spotify, Netflix, YouTube, broadcast, etc.) 

The Science Behind Loudness Measurements

Loudness is a measurement of human perception. If you have not experienced mixing with a loudness meter, you are (hopefully) paying attention to RMS, peak, or VU meters in your DAW or on your hardware. RMS (average level) and peak (loudest level) meters measure levels in decibels relative to full scale (dBFS). The numbers on those meters are based on the voltage of an audio signal. VU meters use a VU scale (where 0 VU is equal to +4 dBu), and like RMS and peak meters, are measuring the voltage of an audio signal.
Those measurements would work to measure loudness – if humans heard all frequencies in the audio spectrum at equal volume levels. But we don’t! Get familiar with the Fletcher-Munson Curve. It is a chart that shows, on average, how sensitive humans are to different frequencies. (Technically speaking, we all hear slightly differently from each other, but this is a solid basis.)

Humans need low frequencies to be cranked up in order to perceive them as the same volume as higher frequencies. And, sound coming from behind us is also weighed louder than sound in front of us. Perhaps it is an instinct that evolved with early humans. As animals, we are still on the lookout for predators that are sneaking up on us from behind.

Instead of measuring loudness in decibels (dB), we measure it in loudness units full scale (LUFS, or interchangeably, LKFS). LUFS measurements account for humans being less sensitive to low frequencies but more sensitive to sounds coming from behind them.

There are a couple more interesting things to know about how loudness meters work. We already mentioned how the EBU standard gates anything below 8 LUFS under the ungated measurement so the really quiet or silent parts do not skew the measurement of the whole mix (which would allow the loudest parts to be way too loud). Loudness standards also dictate the allowed dynamic range of a program (in LUFS). This is important so your audience does not have to tweak the volume to hear people during very quiet scenes, and it saves their ears from getting blasted by a World War Two bomb squadron or a kaiju if they had their stereo turned way up to hear a quiet conversation. (Though every sound designer and mixer knows that there will always be more sensitive listeners who will complain about a loud scene anyway.)

Terms

Here is a list of terms you will see on all loudness meters.

LUFS/LKFS – Loudness Units Full Scale (LKFS = K weighted, but they are effectively the same thing).

Weighting standards – When you mix to a loudness spec in LUFS, also know which standard you should use! The following are the most commonly used standards.

True Peak Max:  Bit of an explanation here. When you play audio in your DAW. you are hearing an analog reconstruction of digital audio data. Depending on how that audio data is decoded, the analog reconstruction might peak beyond the digital waveform. Those peaks are called inter-sample peaks. Inter-sample peaks will not be detected by a limiter or sample peak meter. But a True Peak Meter on a loudness meter will catch them. True peak is measured in dBTP.

Momentary loudness: Loudness at any given moment, for measuring the loudness of a section.

Long-term/ Integrated loudness: This is the average loudness of your mix.

Target Levels: What measurement in LUFS the mix should reach.

Range/LRA: Dynamic range, but in LUFS.  

How To Mix To Loudness Standards

Okay, you know the history, you are armed with the terminology…now what? First, let us talk about the consequences of not mixing to spec.

For every client, there are different devices at the distribution stage that decode your audio and play it out to the airwaves. Those devices have different specifications. The distributor will turn a mix-up or down to normalize the audio to their standards if the mix does not meet specifications. A couple of things happen as a result. One, loss of dynamic range. And, the quietest parts are still too quiet. If there are parts that are too loud, those parts will sound distorted and crushed due to compressed waveforms. The end result is a quiet mix, with no dynamics, with distortion.

To put mixing to loudness in practice, first, start with your ears. Mix what sounds good. Aim for intelligibility and consistency. Keep an eye on your RMS, Peak, or VU meters, but do not worry about LUFS yet.

Your second pass is when you mix to target  LUFS levels. Keep an eye on your loudness meter. I watch the momentary loudness reading because if I am consistently in the ballpark with momentary loudness, I will have a reliable integrated loudness reading and a dynamic range that is not too wide. Limiters can also be used to your advantage.

Then, bounce your mix. Bring the bounce into your session, select the clip, then open your loudness plugin and analyze the bounce. Your loudness plugin will give you a reading with the current specs for your bounce. (Caveat: I am using ProTools terminology. Check if your DAW has a feature similar to AudioSuite.) This also works great for analyzing sections of audio at a time while you are mixing.

Speaking of plugins, here are a few of the most used loudness meters. Insert one of these on your master track to measure your loudness.

Youlean Loudness Meter
This one is top of the list because it is FREE! It also has a cool feature where it shows a linear history of the loudness readings.

iZotope Insight
Insight is really cool. There are a lot of different views, including history and sound field views, and a spectrogram so you can see how different frequencies are being weighted. This plugin measures momentary loudness fast.




Waves WLM Meter

The Waves option may not have a bunch of flashy features like its iZotope competitor, but it does measure everything accurately and comes with an adjustable trim feature. The short-term loudness is accurate but does not bounce around as fast as Insight’s, which I actually prefer.

TC Electronic LMN Meter
I have not personally used this meter, but it looks like a great option for those of us mixing for 5.1 systems. And the radar display is pretty cool!

Wrapping Up: Making Art with Science

The science and history may be a little dry to research, but loudness mixing is an art form itself; Because if listeners have to constantly adjust volume, we are failing at our jobs of creating a distraction and hassle-free experience for our audience. Loudness standards go beyond a set of rules; they are an opportunity for audio engineers to use our scientific prowess to develop our work into a unifying experience.

Resources

First, big thanks to my editors (and fellow audio engineers) Jay Czys and Andie Huether.

The Loudness Standards (Measurement) – LUFS (Cheuks’ Blog)
https://cheuksblog.wordpress.com/2018/04/02/the-loudness-standards-measurement-lufs/#:~:text=Around%202007%2C%20an%20organization%20named,a%20value%20for%20the%20audio.

Loudness: Everything You Need to Know (Production Expert)
https://www.pro-tools-expert.com/production-expert-1/loudness-everything-you-need-to-know

Loud Commercials (The Federal Communications Commission)
https://www.fcc.gov/media/policy/loud-commercials

Loudness vs. True Peak: A Beginner’s Guide (NUGEN Audio)
https://nugenaudio.com/loudness-true-peak/

Worldwide Loudness Standards
https://www.rtw.com/en/blog/worldwide-loudness-delivery-standards.html

Home Recording with Kids in the House

Banished. That’s where my kids would be when I recorded at home pre-pandemic. Not banished to their knowledge, just hanging with their grandparents for at least an entire day, at most overnight. The point is they would be gone, and I’d have sent our dog with them, too, if I’d had my way.

During the pandemic, I found myself itching to record and determined to avoid exposing vulnerable elders to a virus. Short of going into debt building a proper studio at home, I applied my knowledge and resources to develop these strategies:

Directional microphones are your friend. In an ideal recording situation, a large-diaphragm condenser microphone, like my Lauten 220, would be my preference for recording acoustic guitar and vocals. In my loud house, however, noise reduction was key. My kids could be shrieking away in the other room, but with my SM7B pointed away from them, I could sing my heart out with only minimal bleed. Plus, I could point my SM57 or Sennheiser e609 at my amp, again facing away from the children, to get some electric guitar underway.

Sound treatment is your friend. Believe me, I wish my home were properly sound treated – for so many reasons, not just recording purposes. In a pinch, it helps to hang quilts, order some shipping blankets, or hide out in a closet. It all works to cut down on noise. However, be sure to avoid cheap foam options that attenuate only higher frequencies doing little for the low end.

Record directly if you can. If you have a decent DI to connect your instrument to your interface, you won’t have to worry about home interference at all. You can always add effects later, and you may even come up with something innovative and fun with that dry signal. Many interfaces even allow for this without a DI.

Do many takes. While you have everything set up, run through the miked stuff a few times, keeping the same distance from the microphone when you sing or play for consistency’s sake. The little ones won’t scream at the same exact part each time you sing the song unless they’re especially cruel and deliberate about it. You can stitch together the best takes later.

Communicate. Let the kids know you’re going to be doing something special and need quiet for a short while. Talk about what they can be doing to entertain themselves in the meantime. Set boundaries accordingly beforehand and there should be fewer interruptions. Just be prepared to keep your promises if you make any (i.e. ice cream following the recording session, etc.)

It’s never going to be perfect, and of course, it requires flexibility, but it’s completely possible to record at home with your kids around. Breaks may be necessary and you may not get the best sound of your life, but what’s the alternative? Not doing it? Make those perfectly imperfect recordings at home. Lead by example and show the young ones in your life that there is always room for creativity so that they can learn to prioritize their own as they move beyond childhood. And if all else fails, scrap it and try again when they’re asleep.

On Pressing the Button

There are two songs that I remember having written at age six: a rock n’ roll wonder called “Thunder and Lightning” (thunder and lightning/ yeah yeah/ thunder and lightning/ oh yeah) and a narrative style ballad about a little cat that emerged from a mysterious magical flowerbed to become my pet.

I remember them because I recorded them

My dad had a boombox, probably intended for demos of his own. I don’t know what kind it was. All I knew was when I pressed the key with the red circle on it and sang, my voice and songs would come back to me whenever I wanted them to. And I recorded more than those two improvised tunes at age six — I completely commandeered that boombox, interrupted and ruined countless demos of my dad’s. Sometimes I’d prank my younger brother and record the squealing result. Later, in my best robotic voice, I’d make a tape-recorded introduction to my bedroom, to be played whenever somebody rang my “doorbell,” AKA the soundbox from my stuffed tiger, ripped out of its belly and mounted beneath a hand-scrawled sign. I’d even go on to record a faux radio show with my neighborhood friends comprised of comedy bits, original music — vocals only, sung in unison — and, yes, pranking my brother.

Eventually, the boombox either moved on from me, broken or reclaimed by its previous owner, or I moved on from it. I didn’t record anything for a long time, even though I formed other vocals-only bands with friends and continued to write and develop as a songwriter.

Had the little red circle become scary? Was I just a kid, moving from interest to interest, finding myself? Probably the latter. But for some reason, as a young songwriter, I moved from bedroom to bathroom studio, from garage to basement studio, jumping at the chance whenever some dude friend with gear and a designated space offered to get my songs down in some form. Sometimes it was lovely. Other times boundaries were broken and long-term friendships combusted. I persisted because I believed that I needed the help, that I couldn’t record on my own.

Years ago I had a nightmare: I had died without having recorded my music. From a bus full of fellow ghosts with unfinished business, I desperately sang my songs down to the living, hoping someone would catch one, foster it, and let it live. In the early days of the pandemic, this nightmare haunted me. That red circle called to me.

Let’s press that record button, yeah? On whatever we’ve got that has one. I’ve had my songs tracked in spaces sanctioned “studios” by confident men, so why not my own spare room? Why not the record button on my own laptop screen? I’m setting an intention, for myself and for you. When I think about what I wish to provide for you as a Soundgirls blogger, it is this: the permission to record yourself on your own terms, wherever you are in your journey. You are valid.

Reverb Hacks to Make Your Tracks Sparkle

Reverb is a great tool to help bring a bit of life and presence into any track or sound. But why not make it sound even more interesting by applying a plug-in such as an EQ or a compressor to the reverb. It can often give your sound a unique spin as well as being quite fun just to play around with the different sounds that you can achieve.

EQ

The first EQ trick helps with applying reverb to vocals. Have you ever bussed a vocal to a reverb track but still felt like it sounds a bit muddy? Well, try adding an EQ before the reverb on your bus track. Sculpt out the low and high end until you have a rainbow curve. Play around with how much you take out and find what sounds great for your vocals. I often find by doing this you can the clarity of the lyrics as well as achieving a deep, well-echoed sound.  This tip also helps if you’re a bit like me and can’t get enough reverb!

Creating a Pad Sound

If you’re interested in making ambient or classical music, or even pop music that features a soft piano, you might be interested in creating a pad effect. What this does is essentially elongate the sound and sustain it so it gives this nice ambient drone throughout the track.

You can achieve this by creating a bus track and sending your instrument to it. Then open your reverb plugin making sure it is set to 100% wet. You can then play around with setting the decay to around 8.00s to 15.00s. Then send about 60% of your dry instrument track to this bus, making sure to adjust it if it sounds too much. Play around with these settings until you achieve a sound that you like.

In conclusion, Reverb is one of my favourite plugins to play around with and alter. It offers an incredible amount of versatility and can be used in conjunction with many other plugins to create unique and interesting sounds. This can be used on a wide variety of different music genres and comes in handy when you want to add a bit of sparkle to a track.

X