Apple’s Logic Pro X is an indispensable tool in my audio production workflow. After a song idea graduates from Hum, they end up in Logic to get fleshed out as a full demo. Drums, bass, keys, auxiliary guitars, and vocals are added.
I use Logic instead of Pro Tools. Pro Tools feels like Photoshop in many ways. It’s the de facto standard, largely due to its ubiquity after entering its 3rd decade. It’s seen as too big to fail.
Like Photoshop, Pro Tools has never felt native to the macOS experience. It crashes often. Its draconian licensing is relegated to a physical breakout device that has to be purchased separately from the software itself. Its management of takes as playlists feels archaic. Its default set of plugins are paltry in its selection. In the Pro Tools universe, its often easy to feel nickel-and-dimed in order to be productive.
papercuts details matter to a product dork like myself. I’ve tweeted often about my love for Logic. For starters, its price tag is insanely reasonable. It offers incredible plugins that have evolved into an incredibly robust starting point. It doesn’t require any hardware lock-in. It offers a clear path of upgrading from GarageBand and performances with MainStage. It’s my shit. Extremely.
Its best features are baked right into the dough. Its flex time & pitch rival what you could purchase as a third party plugin in the Pro Tools environment. Its quick-swipe comping is an incredible time-saver.
In any session, it’s likely that you’ll record a few takes of vocals. Comping is the act of choosing bits out of each of those takes to build one cohesive performance that’s both in key, on time, and has the right emotion. Comping audio is time consuming in any application. Logic makes it trivial to select the portion, but you’ll still have to listen to each take in earnest to isolate those best bits.
What if this could be simpler? With its Flex Time & Pitch engine, Logic already knows the tempo and pitch data for each of those takes. Allow me to propose the notion of “Automatic comping”. Instead of auditioning each take, what if Logic built a best guess comp based on which notes are closest to the reference note? Here’s what comping looks like currently.
Within the comping dropdown, we could select “Automatic comping (pitch)”. Logic will analyze the notes in each each phrase that was sung, and then comp based on which of those notes vary the least from the reference note.
Here we have a strong starting point based on pitch. In the case of a more rhythmic instrument, bass, or drums, as an example, we could build a comp based on the transients.
Comping is an art. Too many folks get carried away with editing their performances—lining drums up perfectly on the grid, nudging the tuning of a vocal performance, looping a difficult passage. With heavy editing, often these performances lose an undefinable feel. There’s a balancing act there. Of course, one could choose the take that a performer said “was the one” and tune every note and skip comping altogether.
Pitch detection could bring some of the art back to comping. Knowing at a glance which performance is technically correct, we could free up our comping time and allow engineers to focus on the emotion and timbre of the artist’s performance.
At its best, Logic gets out of the way, allowing the artist to focus on expressing their vision for the song. Less time spent on comping means more time experimenting, or hell, writing lyrics worth listening to. Or moving on to another song.
Or maybe robots should just sing this shit in the first place? 🤔
I’m a product designer and musician in Minneapolis, MN. I design and build Hum. I lead a band called The Usual Things. I work at Stack Overflow. I’ve also worked at GitHub and Adobe.