SOUND FORGE Pro Revisited

I’ve reviewed SOUND FORGE a number of times over the years, most recently in 2017. Since its initial development, it has migrated from Sonic Foundry to Sony Creative Software and most recently Magix, a German software developer. Magix’s other products are PC-centric, but SOUND FORGE comes in both Mac and Windows versions.

The updated 3.0 version of SOUND FORGE Pro for the Mac was released in 2017. Although no 4.0 version has been released in the interim, 3.0 was developed as a 64-bit app. Current downloads are, of course, an updated build. Across the product line, there are several versions and bundles, including “lite” SOUND FORGE versions. However, Mac users can only choose between SOUND FORGE Pro Mac 3 or Audio Master Suite Mac. Both include SOUND FORGE Pro Mac, iZotope RX Elements, and iZotope Ozone Elements. The Audio Master Suite Mac adds the Steinberg SpectraLayers Pro 4 analysis/repair application. It’s not listed, but the download also includes the Convrt application, which is an MP3 batch conversion utility.

SOUND FORGE Pro is designed as a dedicated audio mastering application, that does precision audio editing. You can record, edit, and process multichannel audio files (up to 32 tracks) in maximum bit rates of 24-bit, 32-bit, and 64-bit float at up to 192kHz. In addition to the iZotope Elements packages, SOUND FORGE Pro comes with a variety of its own AU plug-ins. Any other AU and VST plug-ins already installed on your system will also show up and work within the application.

Even though SOUND FORGE Pro is essentially a single file editor (as compared with a multi-track DAW, like Pro Tools), you can work with multiple individual files. Multiple files are displayed within the interface as horizontal tabs or in a vertical stack. You can process multiple files at the same time and can copy and paste between them. You can also copy and paste between individual channels within a single multichannel file.

As an audio editor, it’s fast, tactile, and non-destructive, making it ideal for music editing, podcasts, radio interviews, and more. For audio producers, it complies with Red Book Standard CD authoring. The attraction for video editors is its mastering tools, especially loudness control for broadcast compliance. Both Magix’s Wave Hammer and iZotope Ozone Elements’ mastering tools are great for solving loudness issues. That’s aided by accurate LUFS metering. Other cool tools include AutoTrim, which automatically removes gaps of silence at the beginnings and ends of files or from regions within a file.

There is also élastique Timestretch, a processing tool to slow down or speed up audio, while maintaining the correct pitch. Timestretch can be applied to an entire file or simply a section within a file. Effects tools and plug-ins are divided into groups that require processing or those that can be played in real-time. For example, Timestretch is applied as a processing step, whereas a reverb filter would play in real time. Processing is typically fast on any modern desktop or laptop computer, thanks to the application’s 64-bit engine.

Basic editing is as simple as marking a section and hitting the delete key. You can also split a file into events and then trim, delete, move, or copy & paste event blocks. If you slide an event to overlap another, a crossfade is automatically created. You can adjust the fade-in/fade-out slopes of these crossfades.

Even if you already have Logic Pro X, Audition, or Pro Tools installed, SOUND FORGE Pro Mac may still be worth the investment for its simplicity and mastering focus.

©2020 Oliver Peters

ADA Compliance

The Americans with Disabilities Act (ADA) has enriched the lives of many in the disabled community since its introduction in 1990. It affects all of our lives, from wheelchair-friendly ramps on street corners and business entrances to the various accessibility modes in our computers and smart devices. While many editors don’t have to deal directly with the impact of the ADA on media, the law does affect broadcasters and streaming platforms. If you deliver commercials and programs, then your production will be affected in one way or another. Typically the producer is not directly subject to compliance, but the platform is. This means someone has to provide the elements that complete compliance as part of any distribution arrangement, whether it is the producer or the outlet itself.

Two components are involved to meet proper ADA compliance: closed captions and described audio (aka audio descriptions). Captions come in two flavors – open and closed. Open captions or subtitles consists of text “burned” into the image. It is customarily used when a foreign language is spoken in an otherwise English program (or the equivalent in non-English-speaking countries). Closed captions are enclosed in a data stream that can be turned on and off by the viewer, device, or the platform and are intended to make the dialogue accessible to the hearing-impaired. Closed captions are often also turned on in noisy environments, like a TV playing in a gym or a bar.

Audio descriptions are intended to aid the visually-impaired. This is a version of the audio mix with an additional voice-over element. An announcer describes visual information that is not readily obvious from the audio of the program itself. This voice-over fills in the gaps, such as “man climbs to the top of a large hill” or “logos appear on screen.”

Closed captions

Historically post houses and producers have opted to outsource caption creation to companies that specialize in those services. However, modern NLEs enable any editor to handle captions themselves and the increasing enforcement of ADA compliance is now adding to the deliverable requirements for many editors. With this increased demand, using a specialist may become cost prohibitive; therefore, built-in tools are all the more attractive.

There are numerous closed caption standards and various captioning file formats. The most common are .scc (Scenarist), .srt (SubRip), and .vtt (preferred for the web). Captions can be supplied as “embedded” (secondary data within the master file) or as a separate “sidecar” file, which is intended to play in sync with the video file. Not all of these are equal. For example, .scc files (embedded or as sidecar files) support text formatting and positioning, while .srt and .vtt do not. For example, if you have a lower-third name graphic come on screen, you want to move any caption from its usual lower-third, safe-title position to the top of the screen while that name graphic is visible. This way both remain legible. The .scc format supports that, but the other two don’t. The visual appearance of the caption text is a function of the playback hardware or software, so the same captions look different in QuickTime Player versus Switch or VLC. In addition, SubRip (.srt) captions all appear at the bottom, even if you repositioned them to the top, while .vtt captions appear at the top of the screen.

You may prefer to first create a transcription of the dialogue using an outside service, rather than simply typing in the captions from scratch. There are several online resources that automate speech-to-text, including SpeedScriber, Simon Says, Transcriptive, and others. Since AI-based transcription is only as good as the intelligibility of the audio and dialects of the speakers, they all require further text editing/correction through on online tool before they are ready to use.

One service that I’ve used with good results is REV.com, which uses human transcribers for greater accuracy, as well as offering on online text editing tool. The transcription can be downloaded in various formats, including simple text (.txt). Once you have a valid transcription, that file can be converted through a variety of software applications into .srt, .scc, or .vtt files. These in turn can be imported into your preferred NLE for timing, formatting, and positioning adjustments.

Getting the right look

There are guidelines that captioning specialists follow, but some are merely customary and do not affect compliance. For example, upper and lower case text is currently the norm, but you’ll still be OK if your text is all caps. There are also accepted norms when English (or other) subtitles appear on screen, such as for someone speaking in a foreign language. In those cases, no additional closed caption text is used, since the subtitle already provides that information. However, a caption may appear at the top of the screen identifying that a foreign language is being spoken. Likewise, during sections with only music or ambient sounds, a caption may briefly identifying it as such.

When creating captions, you have to understand that readability is key, so the text will not always run perfectly in sync with the dialogue. For instance, when two actors engage in rapid fire dialogue, each caption may stay on longer than the spoken line. You can adjust the timing against that scene so that they eventually catch up once the pace slows down. It’s good to watch a few captioned programs before starting from scratch – just to get a sense of what works and what doesn’t.

If you are creating captions for a program to run on a specific broadcast network or streaming services, then it’s a good idea to find out of they provide a style guide for captions.

Using your NLE to create closed captions

Avid Media Composer, Adobe Premiere Pro, DaVinci Resolve, and Apple Final Cut Pro X all support closed captions. I find FCPX to be the best of this group, because of its extensive editing control over captions and ease of use. This includes text formatting, but also display methods, like pop-on, paint-on, and roll-up effects. Import .scc files for maximum control or extract captions from an existing master, if your media already has embedded caption data. The other three NLEs place the captions onto a single data track (like a video track) within which captions can be edited. Final Cut Pro X places them as a series of connected clips, like any other video clip or graphic. If you perform additional editing, the FCPX magnetic timeline takes care of keeping the captions in sync with the associated dialogue.

Final Cut’s big plus for me is that validation errors are flagged in red. Validation errors occur when caption clips overlap, may be too short for the display method (like a paint-on), are too close to the start of the file, or other errors. It’s easy to find and fix these before exporting the master file.

Deliverables

NLEs support the export of a master file with embedded captions, or “burned” into the video as a subtitle, or the captions exported as a separate sidecar file. Specific format support for embedded captions varies among applications. For example, Premiere Pro – as well as Adobe Media Encoder – will only embed captioning data when you export your sequence or encode a file as a QuickTime-wrapped master file. (I’m running macOS, so there may be other options with Windows.)

On the other hand, Apple Compressor and Final Cut Pro X can encode or export files with embedded captions for formats such as MPEG2 TS, MPEG 2 PS, or MP4. It would be nice if all these NLEs supported the same range of formats, but they don’t. If your goal is a sidecar caption file instead of embedded data, then it’s a far simpler and more reliable process.

Audio descriptions

Compared to closed captions, providing audio description files is relatively easy. These can either be separate audio files – used as sidecar files for secondary audio – or additional tracks on the delivery master. Sometimes it’s a completely separate video file with only this version of the mix. Advanced platforms like Netflix may also require an IMF (Interoperable Master Format) package, which would include an audio description track as part of that package. When audio sidecar files are requested for the web or certain playback platforms, like hotel TV systems, the common deliverable formats are .mp3 or .m4a. The key is that the audio track should be able to run in sync with the rest of the program.

Producing an audio description file doesn’t require any new skills. A voice-over announcer is describing any action that occurs on screen, but which wouldn’t otherwise make sense if you were only listening to audio without that. Think of it like a radio play or podcast version of your TV program. This can be as simple as fitting additional VO into the gaps between actor/host/speaker dialogue. If you have access to the original files (such as a Pro Tools session) or dialogue/music/effects stems, then you have some latitude to adjust audio elements in order to fit in the additional voice-over lines. For example, sometimes the off-camera dialogue may be moved or edited in order to make more space for the VO descriptions. However, on-camera/sync dialogue is left untouched. In that case, some of this audio may be muted or ducked to make space for even longer descriptions.

Some of the same captioning service providers also provide audio description services, using their pool of announcers. Yet, there’s nothing about the process that any producer or editor couldn’t handle themselves. For example, scripting the extra lines, hiring and directing talent, and producing the final mix only require a bit more time added to the schedule, yet permits the most creative control.

ADA compliance has been around since 1990, but hasn’t been widely enforced outside of broadcast. That’s changing and there are no more excuses with the new NLE tools. It’s become easier than ever for any editor or producer to make sure they can provide the proper elements to touch every potential viewer.

For additional information, consult the FCC guidelines on closed captions.

The article was originally written for Pro Video Coalition.

©2020 Oliver Peters

Accusonus ERA4

It’s fair to say that most video editors, podcast producers, and audio enthusiasts don’t possess the level of intimate understanding of audio filters that a seasoned recording engineer does. Yet each still has the need to present the cleanest, most professional-sounding audio as part of their productions. Some software developers have sought to serve this market with plug-ins that combine the controls into a single knob or slider. The first of these was Waves Audio with their OneKnob series. Another company using the single-knob approach is a relative newcomer, Accusonus.

I first became aware of Accusonus through Avid. Media Composer license owners have been offered loyalty add-ons, such as plug-ins, which recently included the Accusonus ERA3 Voice Leveler plug-in. I found this to be a very useful tool and so when Accusonus offered to send the new ERA4 Bundle for my evaluation, I was more than happy to give the rest of the package a test run. ERA4 was released at the end of June in a Standard and Pro bundle along with discounted, introductory pricing, available until the end of July. You may also purchase each of these filters individually.

Audio plug-ins typically come in one of four formats: AU (Mac only), VST (Mac or Windows), VST3 (Mac or Windows), and AAX (for Avid Media Composer and Pro Tools). When you purchase audio filters, they don’t necessarily come in all flavors. Sometimes, plug-ins will be AU and VST/VST3, but leave out AAX. Or will only be for AAX. Or only AU. Accusonic plug-ins are installed as all four types on a Mac, which means that a single purchase covers most common DAWs and NLEs (check their site for supported hosts). For example, my Macs include Final Cut Pro X, Logic Pro X, Audition, Premiere Pro, and Media Composer. The ERA4 plug-ins work in all of these.

I ran into some issues with Resolve. The plug-ins worked fine on the Fairlight page of Resolve 16 Studio Beta. That’s on my home machine. However, the Macs at work are running the Mac App Store version of Resolve 15 Studio. There, only the VST versions could be applied and I had to re-enter each filter’s activation code and relaunch. I would conclude from this that Resolve is fine as a host, although there may be some conflicts in the Mac App Store version. That’s likely due to some differences between it and the software you download directly from Blackmagic Design.

Another benefit is that Accusonus permits each license key to be used on up to three machines. If a user has both a laptop and a desktop computer, the plug-in can be installed and activated on each without the need to swap authorizations through an online license server or move an iLok dongle between machines. The ERA4 installers include all of the tools in the bundle, even if you only purchased one. You can ignore the others, uninstall them, or test them out in a trial mode. The complete bundle is available and fully functional for a 14-day free trial.

ERA4 Bundles

I mentioned the Waves One Knob filters at the top, but there’s actually little overlap between these two offerings. The One Knob series is focused on EQ and compression tasks, whereas the ERA4 effects are designed for audio repair. As such, they fill a similar need as the iZotope’s RX series.

The ERA4 Standard bundle includes six audio plug-ins: Noise, Reverb, and Plosive Removers, De-Esser, De-Clipper, and the Voice Leveler. The Pro bundle adds two more: the more comprehensive De-Esser Pro and ERA-D, which is a combined noise and reverb filter for more advanced processing than the two individual filters. If you primarily work with well-recorded studio voice-overs or location dialogue, then most likely the Standard bundle will be all you need. However, the two extra filters in the Pro bundle come in handy with more problematic audio. Even productions with high values occasionally get stuck with recordings done in challenging environments and last-minute VOs done on iPhones. It’s certainly worth checking out the full package as a trial.

While Accusonus does use a single-control approach, but it’s a bit simplistic to say that you are tied to only one control knob. Some of the plug-ins do offer more depth so you can tailor your settings.  For instance, the Noise Remover filter offers five preset curves to determine the frequencies that are affected. Each filter includes additional controls for the task at hand.

In use

Accusonus ERA4 filters are designed to be easy to use and work well in real-time. When all I need to do is improve audio that isn’t a basket case, then the ERA filters at their default settings do a wonderful job. For example, a VO recording might require a combination of Voice Leveler (smooth out dynamics), De-Esser (reduce sibilance), and Plosive Remover (clean up popping “p” sounds). Using the default control level (40%) or even backing off a little improved the sound.

It was the more problematic audio where ERA4 was good, but not necessarily always the best tool. In one case I tested a very, heavily clipped VO recording. When I used ERA4 De-Clipper in Final Cut Pro X, I was able to get similar results to the same tool from iZotope RX6. However, doing the same comparison in Audition yielded different results. Audition is designed to preview an effect and then apply it. The RX plug-in at its extreme setting crackled in real-time playback, but yielded superior results compared with the ERA4 De-Clipper after the effect was applied (rendered). Unfortunately FCPX has no equivalent “apply,” “render and replace,” or audio bounce function, so audio has to stay real-time, which gives Accusonus a performance edge in FCPX. For most standard audio repair tasks, Accusonus’ plug-ins were equal or better than most other options, especially those built into the host application.

I started out talking about the Voice Leveler plug-in, because that’s an audio function I perform often, especially with voice-overs. It helps to make the VO stand out in the mix against music and sound effects. This is an intelligent compressor, which means it tries to bring up all audio and then compress peaks over a threshold. But learn the controls before diving in. For example, it includes a breath control. Engaging this will prevent the audio from pumping up in volume each time the announcer takes a breath. As with all of the ERA4 filters, there is a small, scrolling waveform in the plug-in’s control panel. Areas that were adjusted by the filter are highlighted, so you can see when it is active.

Voice Leveler is a good VO tool, but that type is one of the more subjective audio filters. Some editors or audio engineers compress, some limit, and others prefer to adjust levels only manually. My all-time favorite is Wave’s Vocal Rider. Unlike a compressor, it dynamically raises and lowers audio levels between two target points. To my ears, this method yields a more open sound than heavy compression. But when its normal MSRP is pretty expensive. I also like the Logic Pro X Compressor, which is available in Final Cut Pro X. It mimics various vintage compressors, like Focusrite Red or the DBX 160X. I feel that it’s one of the nicest sounding compressors, but is only available in the Apple pro apps. Adobe users – you are out of luck on that one.

From my point-of-view, the more tools the better. You never know when you might need one. The Accusonus ERA4 bundle offers a great toolset for anyone who has to turn around a good-sounding mix quickly. Each bundle is easy to install and activate and even easier to use. Operation is real-time, even when you stack several together. Accusonus’ current introductory price for the bundles is about what some individual plug-ins cost from competing companies, plus the 14-day trial is a great way to check them out. If you need to build up your audio toolbox, this is a solid set to start out with.

Check out Accusonus’ blog for tips on using the ERA plug-ins.

©2019 Oliver Peters

Luca Visual FX builds Mystery & Suspense

For most editors, creating custom music scores tends to fall into the “above my pay grade” category. If you are a whizz with GarageBand or Logic Pro X, then you might dip into Apple’s loop resources. But most commercials and corporate videos are easily serviced by the myriad of stock music sites, like Premium Beat and Music Bed. Some music customization is also possible with tracks from companies like SmartSound.

Yet, none of the go-to music library sites offer curated, genre-based, packages of tracks and elements that make it easy to build up a functional score for longer dramatic productions. Such projects are usually the work of composers or a specific music supervisor, sound designer, or music editor doing a lot of searching and piecing together from a wide range of resources.

Enter Luca Visual FX – a developer best known for visual effects plug-ins, such as Light Kit 2.0. It turns out that Luca Bonomo is also a composer. The first offering is the Mystery & Suspense Music and Sound Library, which is a collection of 500 clips, comprising music themes, atmospheres, drones, loops, and sound effects. This is a complete toolkit designed to make it easy to combine elements, in order to create a custom score for dramatic productions in the mystery or suspense genre.

These tracks are available for purchase as a single library through the LucaVFX website. They are downloaded as uncompressed, stereo AIF files in a 24-bit/96kHz resolution. This means they are of top quality and compatible with any Mac or PC NLE or DAW application. Best yet, is the awesome price of $79. The package is licensed for a single user and may be used for any audio or video production, including for commercial purposes.

Thanks to LucaVFX, I was able to download and test out the Library on a recent short film. The story is a suspense drama in the style of a Twilight Zone episode, so creating a non-specific, ethereal score fits perfectly. Drones, dissonance, and other suspenseful sounds are completely in line, which is where this collection shines.

Although I could have used any application to build this, I opted for Apple’s Final Cut Pro X. Because of its unique keyword structure, it made sense to first set up a separate FCPX library for only the Mystery & Suspense package. During import, I let FCPX create keyword collections based on the Finder folders. This keeps the Mystery & Suspense FCPX library organized in the same way as they are originally grouped. Doing so, facilitates fast and easy sorting and previewing of any of the 500 clips within the music library. Then I created a separate FCPX library for the production itself. With both FCPX libraries open, I could quickly preview and place clips from my music library to the edit sequence for the film, located within the other FCPX library.

Final Cut uses Connected Clips instead of tracks. This means that you can quickly build up and align overlapping atmospheres, transitions, loops, and themes for a densely layered music score in a very freeform manner. I was able to build up a convincing score for a half-hour-long piece in less that an afternoon. Granted, this isn’t mixed yet, but at least I now have the musical elements that I want and where I want them. I feel that style of working is definitely faster in Final Cut Pro X – and more conducive to creative experimentation – but it would certain work just as well in other applications.

The Mystery & Suspense Library is definitely a winner, although I do have a few minor quibbles. First, the music and effects are in keeping with the genre, but don’t go beyond it. When creating a score for this kind of production, you also need some “normal” or “lighter” moods for certain scenes or transitions. I felt that was missing and I would still have to step outside of this package to complete the score. Secondly, many of the clips have a synthesized or electronic tone to them, thanks to the instruments used to create the music. That’s not out of character with the genre, but I still would have liked some of these to include more natural instruments than they do. In fairness to LucaVFX, if the Mystery & Suspense Library is successful, then the company will create more libraries in other genres, including lighter fare.

In conclusion, this is a high quality library perfectly in keeping with its intended genre. Using it is fast and flexible, making it possible for even the most musically-challenged editor to develop a convincing, custom score without breaking the bank.

©2018 Oliver Peters

Audio Mixing with Premiere Pro

When budgets permit and project needs dictate, I will send my mixes out-of-house to one of a few regular mixers. Typically that means sending them an OMF or AAF to mix in Pro Tools. Then I get the mix and split-tracks back, drop them into my Premiere Pro timeline, and generate master files.

On the other hand, a lot of my work is cutting simple commercials and corporate presentations for in-house use or the web, and these are often less demanding  – 2 to 8 tracks of dialogue, limited sound effects, and music. It’s easy to do the mix inside of the NLE. Bear in mind that I can – and often have – done such a mix in Apple Logic Pro X or Adobe Audition, but the tools inside Premiere Pro are solid enough that I often just keep everything – mix included – inside my editing application. Let’s walk though that process.

Dealing with multiple channels on source clips

Start with your camera files or double-system audio recordings. Depending on the camera model, Premiere Pro will see these source clips as having either stereo (e.g. a Canon C100) or multi-channel mono (e.g. ARRI Alexa) channels. If you recorded a boom mic on channel 1 and a lavaliere mic on channel 2, then these will drop onto your stereo timeline either as two separate mono tracks (Alexa) – or as a single stereo track (C100), with the boom coming out of the left speaker and the lav out of the right. Which one it is will strictly depend on the device used to generate the original recordings.

First, when dual-mic recordings appear as stereo, you have to understand how Premiere Pro deals with stereo sources. Panning in Premiere Pro doesn’t “shift” the audio left, right, or center. Instead, it increases or decreases the relative volume of the left or right half of this stereo field. In our dual-mic scenario, panning the clip or track full left means that we only hear the boom coming out of the left speaker, but nothing out of the right. There are two ways to fix this – either by changing the channel configuration of the source in the browser – or by changing it after the fact in the timeline. Browser changes will not alter the configuration of clips already edited to the timeline. You can change one or more source clips from stereo to dual-mono in the browser, but you can’t make that same type of change to a clip already in your sequence.

Let’s assume that you aren’t going to make any browser changes and instead just want to work in your sequence. If your source clip is treated as dual-mono, then the boom and lav will cut over to track 1 and 2 of your sequence – and the sound will be summed in mono on the output to your speaks. However, if the clip is treated as stereo, then it will only cut over to track 1 of your sequence – and the sound will stay left and right on the output to your speakers. When it’s dual-mono, you can listen to one track versus the other, determine which mic sounds the best, and disable the clip with the other mic. Or you can blend the two using clip volume levels.

If the source clip ends up in the sequence as a stereo clip, then you will want to determine which one of the two mics you want to use for the best sound. To pick only one mic, you will need to change the clip’s audio configuration. When you do that, it’s still a stereo clip, however, both “sides” can be supplied by either one of the two source channels. So, both left and right output will either be the boom or the lav, but not both. If you want to blend both mics together, then you will need to duplicate (option-drag) the audio clip onto an adjacent timeline track, and change the audio channel configuration for both clips. One would be set to the boom for both channels and the other set to only the lav for its two channels. Then adjust clip volume for the two timeline clips.

Configuring your timeline

Like most editors, while I’m working through the stages of rough cutting on the way to an approved final copy, I will have a somewhat messy timeline. I may have multiple music cues on several tracks with only one enabled – just so I can preview alternates for the client. I will have multiple dialogue clips on a few tracks with some disabled, depending on microphone or take options. But when I’m ready to move to the finishing stage, I will duplicate that sequence to create a “final version” and clean that one up. This means getting rid of any disabled clips, collapsing my audio and video clips to the fewest number of tracks, and using Premiere’s track creation/deletion feature to delete all empty tracks – all so I can have the least amount of visual clutter. 

In other blog posts, I’ve discussed working with additional submix buses to create split-track exports; but, for most of these smaller jobs, I will only add one submix bus. (I will explain its purpose in a moment.) Once created, you will need to open the track mixer panel and route the timeline channels from the master to the submix bus and then the output of the submix bus back to the master.

Plug-ins

Premiere Pro CC comes with a nice set of audio plug-ins, which can be augmented with plenty of third-party audio effects filters. I am partial to Waves and iZotope, but these aren’t essential. However, there are several that I do use quite frequently. These three third-party filters will help improve any vocal-heavy piece.

The first two are Vocal Rider and MV2 from Waves and are designed specifically for vocal performances, like voice-overs and interviews. These can be pricey, but Waves has frequent sales, so I was able to pick these up for a fraction of their retail price. Vocal Rider is a real-time, automatic volume adjustment tool. Set the bottom and top parameters and let Vocal Rider do the rest, by automatically pushing the volume up or down on-the-fly. MV2 is similar, but it achieves this through compression on the top and bottom ends of the range. While they operate in a similar fashion, they do produce a different sound. I tend to pick MV2 for voice-overs and Vocal Rider for interviews.

We all know location audio isn’t perfect, which is where my third filter comes in. FxFactory is knows primarily for video plug-ins, but their partnership with Crumplepop has added a nice set of audio filters to their catalog. I find AudioDenoise to be quite helpful and fast in fixing annoying location sounds, like background air conditioning noise. It’s real-time and good-sounding, but like all audio noise reduction, you have to be careful not to overdo it, or everything will sound like it’s underwater.

For my other mix needs, I’ll stick to Premiere’s built-in effects, like EQ, compressors, etc. One that’s useful for music is the stereo imager. If you have a music cue that sounds too monaural, this will let you “expand” the track’s stereo signal so that it is spread more left and right. This often helps when you want the voice-over to cut through the mix a bit better. 

My last plug-in is a broadcast limiter that is placed onto the master bus. I will adjust this tight with a hard limit for broadcast delivery, but much higher (louder allowed) for web files. Be aware that Premiere’s plug-in architecture allows you to have the filter take affect either pre or post-fader. In the case of the master bus, this will also affect the VU display. In other words, if you place a limiter post-fader, then the result will be heard, but not visible through the levels displayed on the VU meters.

Mixing

I have used different mixing strategies over the years with Premiere Pro. I like using the write function of the track mixer to write fader automation. However, I have lately stopped using it – instead going back to manual keyframes within the clips. The reason is probably that my projects tend to get revised often in ways that change timing. Since track automation is based on absolute timeline position, keyframes don’t move when a clip is shifted, like they would when clip-based volume keyframes are used.

Likewise, Adobe has recently added Audition’s ducking for music to Premiere Pro. This uses Adobe’s Sensei artificial intelligence. Unfortunately I don’t find to be “intelligent” enough. Although sometimes it can provide a starting point. For me, it’s simply too coarse and doesn’t intelligently adjust for areas within a music clip that swell or change volume internally. Therefore, I stick with minor manual adjustments to compensate for music changes and to make the vocal parts easy to understand in the mix. Then I will use the track mixer to set overall levels for each track to get the right balance of voice, sound effects, and music.

Once I have a decent balance to my ears, I will temporarily drop the TC Electronic (included with Premiere Pro) Radar loudness plug-in to make sure my mix is CALM-compliant. This is where the submix bus comes in. If I like the overall balance, but I need to bring everything down, it’s an easy matter to simply lower the submix level and remeasure.

Likewise, it’s customary to deliver web versions with louder volume levels than the broadcast mix. Again the submix bus will help, because you cannot raise the volume on the master – only lower it. If you simply want to raise the overall volume of the broadcast mix for web delivery, simply raise the submix fader. Note that when I say louder, I’m NOT talking about slamming the VUs all the way to the top. Typically, a mix that hits -6 is plenty loud for the web. So, for web delivery, I will set a hard limit at -6, but adjust the mix for an average of about -10.

Hopefully this short explanation has provided some insight into mixing within Premiere Pro and will help you make sure that your next project sounds great.

©2018 Oliver Peters