Melodyne can do much more than vocal pitch correction. Previous tips have covered how to do envelope-controlled flanging and polyphonic guitar-to-MIDI conversion with Melodyne Essential. The following techniques add mind-bending effects to lead guitar, and don’t involve pitch correction per se.
However, all three tips require the Pitch Modulation and Pitch Drift tools. These tools are available only in versions above Essential (Assistant, Editor, and Studio). A Melodyne 5 trial version incorporates these versions. The trial period is 30 days, during which there are no operational limitations.
Fig. 1 shows a lead guitar line, before and after processing, that incorporates all three processing techniques. This is the melody line used in the before-and-after audio example.
Vibrato Emphasis/De-Emphasis
To increase or decrease the vibrato amount, use Melodyne’s Pitch Modulation tool. In fig. 1, the increased vibrato effect is most visually obvious at the end of measures 43, 49, and 52. To change the amount of vibrato, click on the blob with the Pitch Modulation tool. While holding the mouse button down, drag up (more vibrato) or down (less vibrato). Extreme vibrato amounts can sound like whammy bar-based vibrato.
Synthesized Slides
Notes with moderate bending can bend pitch up or down over as much as several semitones:
1. Select the Pitch Modulation tool.
2. Click on a blob that incorporates a moderate bend.
3. While holding the mouse button down, drag up to increase the bend up range. With most notes, you can invert the bend by dragging down.
Because a synthesized bend can cover a wider pitch range than physical strings, this effect sounds like you’re using a slide on your finger. In fig. 1, see measures 44 and 45 for examples of upward bends. The end of measure 47 shows an increased downward bend.
For the most predictable results, the note you want to bend should:
Slide Up to Pitch, or Slide Down to Pitch
This is an unpredictable technique, but when it works, note transitions acquire a “smooth” character. In fig. 1, note the difference between the modified and unmodified pitch slides in measures 43, 44, 47, 49, 50, and 51.
To add this kind of slide, click on a blob with the Pitch Drift tool, and then drag up or down. The slide’s character depends on what happens during, before, and after the note. Sometimes using Pitch Modulation to initiate a slide, and Pitch Drift to modify the slide further, works best. Sometimes the reverse is true.
This is a trial-and-error process. With experience, you’ll be able to recognize which blobs are good candidates for slides.
The ”Hearing is Believing” Audio Examples
Guitar Solo.mp3 is an isolated guitar solo that uses none of these techniques.
Guitar Solo with Melodyne.mp3 uses all of these techniques on various notes.
To hear these techniques in a musical context that shows how they can add a surprising amount of emotion to a track,this link takes you to a guitar solo in one of my songs on YouTube. The solo uses all three effects.
Synchronized echo effects, particularly dotted eighth-note delays (i.e., intervals of three 16th notes), are common in EDM and dance music productions. The following audio example applies this type of Analog Delay effect to Mai Tai.
However, you can also create echoes for virtual instruments by copying, offsetting, and editing an instrument track’s MIDI note data. In the next audio example, the same MIDI track has been copied and delayed by an 8th note (fig. 1).
In the original instrument track, the MIDI note data velocities are around 127. These values push the filter cutoff close to maximum. Reducing the copied Echo track’s velocity to about half creates notes that don’t open up the Mai Tai’s filter as much. So, the MIDI-generated delay’s timbre is different.
The next track copies the original track again. But this time, the notes are transposed up an octave, and offset by a dotted eighth-note compared to the original (fig. 2). For this track, the velocities are about halfway from the maximum velocity.
Now we have our “MIDI-accelerated” echo effect for the virtual instrument’s notes, which are still processed by the original Analog Delay effect. The final audio example highlights how the combination of analog delay and MIDI note delay evolves over 8 measures.
These audio examples are only the start of what you can do with MIDI echo by offsetting, transposing, and altering note velocities. You can even feed the different MIDI tracks into different virtual instruments, and create amazing polyrhythms. But why stop there? Hopefully, this blog post will inspire you to come up with your own signature variations on this technique.
This is a follow-up to the Friday Tips in the Real World blog post that appeared in 2020. It was well-received, so I figured it was time for an update.
Although many of the Friday Tips include an audio example of how a tip affects the sound, that’s different from the real-world context of a musical production. So, this blog post highlights how selected tips were used in my recent music/video project, Unconstrained. The project was recorded, mixed, and mastered entirely in Studio One.
As you might expect, the workflow-related tips were used throughout, as were some of the audio tips. For example, all the vocals used the Better Vocals with Phrase-by-Phrase Normalization technique, and all the guitar parts followed the Amp Sims: Garbage In, Garbage Out tip. The Chords Track and Project Page updates were crucial to the entire project as well.
The links take you to specific parts of the songs that showcase the tips, accompanied by links to the relevant Friday Tip blog post. If you’re curious about specific production techniques used in the project, whether they’re included in this post or not, feel free to ask questions in the comments section.
One reader’s comment for the Lead Guitar Editing Hack blog post mentioned how useful this technique is. If you missed it the first time, here’s what it sounds like applied to a solo. Attenuating the attack gives the melody line a synth-like, otherworldly sound. Incidentally, if you back up the video to 6:20, the cello sounds are from Presence. I tried the “industry standard” orchestra programs, but I liked the Presence cellos more. Also, I used the “time trap” technique from the Fun with Tempo Tracks post to slow the cellos slightly before going full tilt into the solo.
The Magic Stereo blog post described a novel way to add motion to rhythm parts, like piano and guitar. This excerpt uses that technique to move the guitar in stereo, but without conventional panning. Later on in the song, the drums use Harmonic Editing to give a sense of pitch. The post Melodify Your Beats describes this technique. But in this song, the white noise wasn’t needed because the drums had enough of a frequency range so that harmonic editing worked well.
I wrote about the EDM-style “pumping” effect in the post “Pump” Your Pads and Power Chords, and it goes most of the way through this song. The reason why I chose this section is because the solo uses Presence, which I think may be underrated by some people.
Another topic that’s dear to my heart is blues harmonica, and it loves distortion—as described in the blog post Blues Harmonic FX Chain. It’s wonderful how amp sims can turn the thin, reedy sound of a harmonica into something with so much power it’s almost like a brass section. However, note that this example uses a revised version of the original FX Chain, based on Ampire. (The revised version is described in The Huge Book of Studio One Tips and Tricks.)
The blog post Studio One’s Session Bass Player generated a lot of comments. But does the technique really work? Well, listen to this example and decide for yourself. I needed a scratch bass part but it ended up being so much like what I wanted that I made only a couple tweaks…done. For a guitar solo in the same song, I tried a bunch of wah pedals but the one that worked best was Ampire’s.
I still think Studio One’s ability to do polyphonic MIDI guitar courtesy of Melodyne (even the Essential version) is underrated. This “keyboard” part uses Mai Tai driven by MIDI guitar. The MIDI part was derived from the guitar track that’s doubling the Mai Tai. For more information, see the blog post Melodyne Essential = Polyphonic MIDI Guitar. Incidentally, except for the sampled bass and choir, all the keyboard sounds were from Mai Tai. If you’ve mostly been using third-party synths, spend some time re-acquainting yourself with Mai Tai and Presence. They can really deliver.
As the post Synthesize OpenAIR Reverb Impulses in Studio One showed, it’s easy to create your own reverb impulses for OpenAIR. In this excerpt, the female background vocals, male harmony, and harmonica solo all used impulses I created using this technique. (The only ambience on the lead vocal was the Analog Delay). Custom impulses are also used throughout Vortex and the subsequent song, What Really Matters (which also uses the Lead Guitar Hack for the solo).
I’m just getting started with my project for 2023, and it’s already generating some new tips that you’ll be seeing in the weeks ahead. I hope you find them helpful! Meanwhile, here’s the link to the complete Unconstrained project.
Announcement: Version 1.4.1 of The Huge Book of Studio One Tips and Tricks is a free update to owners of previous versions. Simply download the book again from your PreSonus account, and it will be the most recent version. This is a “hotfix” update for improved compatibility with Adobe Acrobat’s navigation functions. The content is the same as version 1.4.
Studio One has a solid repertoire of EQs: the Fat Channel EQs, Pro EQ3, Ampire’s Graphic Equalizer, and the Autofilter. Even the Multiband Dynamics can serve as a hip graphic EQ. With this wealth of EQs, it’s potentially easy to overlook the Channel Strip’s EQ section (fig. 1). Yet it’s significantly different from the other EQs.
Back to the 60s
In the late 60s, Saul Walker (API’s founder) introduced the concept of “proportional Q” in API equalizers. To this day, engineers praise API equalizers for their “musical” sound, and much of this relates to proportional Q.
The theory is simple. At lower gains, the bandwidth is wider. At higher gains, it becomes narrower. This is consistent with how we often use EQ. Lower gain settings are common for tone-shaping. Increasing the gain likely means you want a more focused effect.
The concept works similarly for cutting. If you’re applying a deep cut, you probably want to solve a problem. A broad cut is more about shaping tone. Also, because cutting mirrors the response of boosting, proportional Q equalizers make it easy to “undo” equalization settings. For example, if you boosted drums around 2 kHz by 6 dB, cutting by 3 dB produces the same curve as if the drums had originally been boosted by 3 dB.
Proportional Q also works well with vocals and automation. For less intense parts, add a little gain in the 2-4 kHz range. When the vocal needs to cut through, raise the gain to increase the resonance and articulation.
The Channel Strip’s Adapt Q button converts the response to proportional Q. Let’s look at how proportional Q affects the various responses.
High and Low Stages
The High band is a 12 dB/octave high shelf. In fig. 2, the Boost is +15 dB, at a frequency of 2.5 kHz. The top image shows the EQ without Adapt Q. The lower image engages Adapt Q, which increases the shelf’s resonance.
Fig. 3 shows what happens with 6 dB of gain. With Adapt Q enabled, the Q is actually less than the corresponding amount of Q without Adapt Q.
Cutting flips the curve vertically, but the shape is the same. With the Low shelf filter, the response is the mirror image of the High shelf.
Midrange Stage
The Midrange EQ stage has variable Gain and Frequency. There’s no Q control, but the filter works with Adapt Q to increase Q with more gain or cut (fig. 4).
With 6 dB Gain, the Q is essentially the same, regardless of the Adapt Q setting (fig. 5).
Finally, another Adapt Q characteristic is that the midrange section’s slope down from either side of the peak (called the “skirt”) hits the minimum amount of gain at the same upper and lower frequencies, regardless of the gain. This is different from a traditional EQ like the Pro EQ3, where the skirt narrows with more Q (fig. 6).
Perhaps best of all, the Channel Strip draws very little CPU power. So, if you need more stages of EQ, go ahead and insert several Channel Strips in series, or in parallel using a Splitter or buses. And don’t forget—the Channel Strip also has dynamics 😊!
A single Automation track can set up a session’s preamp levels and phantom power in the Quantum interface, as well as the older Studio 192. So, you can stop taking the time to reset preamp levels if you do lots of different sessions—let Studio One set up the preamps whenever you call up a specific song.
For example, I mostly use three vocal mics. However, their optimum gain settings vary for narration, music vocals, or recording my main background vocalist—who needs different gain settings depending on whether she’s doing upfront vocals, or ooohs/ahhhs.
To call up specific preamp levels for different songs, simply create an automation track (or tracks) at the song’s beginning. Then, when you first hit record, the track sets up levels and (with Quantum and Studio 192) phantom power on/off for up to 8 channels. The next time you call up that song, the mics will be at the right levels, with phantom power set as desired. Here’s how to do it.
1. In Universal Control, under MIDI Control, select Internal. Or, choose Enabled if you also want to be able to control Quantum from an external controller.
2. Choose Studio One > Options (Windows) or Preferences (macOS), and select External Devices.
3. Select Add. Choose New Instrument.
4. For Receive From, choose Quantum MIDI In. For Send To, choose Quantum MIDI Out. Also tick MIDI Channels 1 – 8 (fig. 1). Then, click OK.
5. Create an Automation track. To show automation, type keyboard shortcut A.
6. From the track’s automation drop-down menu, click on Add/Remove. In the right pane, unfold the External Devices folder, then unfold the folder for the Channel that corresponds to the preamp you want to control (Channel 1 = Preamp 1, Channel 2 = Preamp 2…Channel 8 = Preamp 8). See fig. 2.
7. For each preamp you want to set up, add CC7 (this controls preamp volume) and CC14 (controls phantom power). For example, I typically set up channels 6, 7, and 8. After adding these continuous controllers, the Automation menu’s right side looks like fig. 3 (except with your specific channel numbers). Click on Close after making your selections.
8. To set up the preamp levels, choose the parameter you want to program in the Automation track’s drop-down menu. Note that in the documentation, the phantom power control settings are reversed. The correct values for CC14 are 0 to 63 = Off, and 64 to 127 = On. To set the preamp level, with Universal Control open, adjust the envelope for the desired preamp gain reading (you can also see the level on the Quantum’s display).
9. Set the initial level and phantom power parameter values for the chosen preamps. Now your automation track will reproduce those settings, exactly as programmed, the next time you open the song. Given that I do voiceover or narration for at least one video a week, I can’t tell you how much time that saves—I load my narration template, and don’t even have to think about adjusting levels before hitting record.
Getting Fancy
A cool trick is to reserve a song’s first measure for doing the setup. Turn on phantom power at the song start, but fade up the volume to the desired level after the phantom power is on. That avoids power-on spikes from the mics. However, when adjusting the level envelope, the preamp knobs change only if you adjust the left-most node. So, set the preamp level you want with this node, then move it to the right on the timeline. Create another node at 0 that fades up to the node you moved, which sets the final volume.
Another trick is to have more than one automation track. For example, on most songs I have a setup track for me, and a setup track for the background singer. When she does overdubs, I turn my automation track to Off, and set her automation track to Read so it sets her levels.
Coda: Windows Meets Thunderbolt
The first time I tried Quantum on a Mac, it worked perfectly. With Windows, well…it’s Windows. My computer is a PC Audio Labs Rok Box (great machine, by the way) with dual Thunderbolt 3 ports. I used Apple’s TB3-to-TB2 adapter—no go. I found a new Thunderbolt driver for the motherboard, and asked PC Audio Labs tech support about whether I should install it. They advised doing so, and said if I had problems, they’d bail me out. But after installation, the Quantum’s power button’s color was still Unhappy Red instead of Happy Blue. I was about to contact support again, but stumbled on a program in the computer called Thunderbolt Control Center. I opened it, which showed Quantum was connected—but I hadn’t given the computer “permission” to connect. So, I gave permission. With its new-found freedom, the Quantum burst into its low-latency glory.
The moral of the story: Thunderbolt has many variables on Windows than macOS. But as with life itself, perseverance furthers.
Studio One’s chorus gives the “wet” sound associated with chorus effects. But I wanted a chorus that went beyond wet to drenched—something that could swirl in the background of a thick arrangement, and shower the stereo field like a sprinkler system. Check out the audio example: the first part is the Drenched Chorus, and the second part is the standard Chorus.
This FX Chain for Studio One Professional (see the download link at the end) supercharges the wet sound by inserting two choruses in parallel, and reversing the phase for one of them (fig. 1). This cancels any dry sound, leaving only the animation from the stereo chorus. A Mixtool provides the phase reversal. Another Mixtool at the input adds gain, to compensate for the level that’s lost through phase cancellation.
The Choruses are essentially set to the default preset, but with Low Freq at minimum and High Freq at maximum for the widest frequency response. Fig. 2 uses a phase meter to compare the Drenched Chorus (left) with the standard Chorus (right), playing the same section of a guitar part. Note how the Drenched Chorus puts a lot of energy into the sides, which accounts for the big stereo image. Meanwhile, the center has less level than the standard chorus, due to any vestiges of dry signal being removed.
This effect is designed for stereo playback, but note that in fig. 2, the Drenched Chorus’s correlation is negative. Normally you want to avoid this, because audio with negative correlation will cancel when played back in mono. However, the correlation swings wildly between positive and negative, so it’s not much of an issue. With mono playback, all that happens is a slight level loss due to occasional negative correlations. The effect still sounds like a chorus, although of course you lose the cool stereo effects.
How to Use It
Download the FX Chain, and drag it into a channel’s insert. The Macro Controls (fig. 3) affect only Chorus 2. Here’s what they do:
Download the FX Chain below!
Studio One offers several ways to “de-ess” excessive sibilants (“s” sounds). De-essing combines compression and EQ. The EQ focuses on the frequency range where sibilants are most prominent. Then, compression reduces this range’s level when sibilants are present. Prior to version 6, using Multiband Dynamics was the best way to do de-essing.
With version 6, the Pro EQ3’s new dynamic EQ functionality is excellent for reducing ess sounds. However, the equally new De-Esser (fig. 1) is designed specifically for the job of fixing excessive sibilance, quickly and effectively.
To get the most out of the De-Esser, note that some of the parameters work together as a team. So, alternating edits between some controls is often the best way to optimize the effect.
Getting Started
Ess sounds tend to be short. By the time you’ve started to tweak a parameter, the ess sound has likely already ended. So, for easy editing, create a short loop on a word with the prominent ess sound.
Frequency and Listen
1. Enable Listen.
2. Vary Frequency until you find the frequency where the ess sound is most prominent.
Solo and S-Reduction
3. After identifying the frequency, enable Solo. You’ll hear only the ess sound whose volume is being reduced.
4. Adjust S-Reduction to get a feel for the optimum ess reduction amount. At 0.00 dB, there’s no reduction. At ‑60.00, sounds in addition to the ess sound will likely be reduced.
5. Next, turn off Solo, and adjust S-Reduction in context with the looped word. Less negative S-Reduction values concentrate on reducing the ess sound’s initial transient. More negative values reduce more of the ess sound past the initial transient.
6. Do a final Frequency parameter check to optimize the high-frequency response with de-essing. For example, you may be able to raise the frequency to retain more highs, yet still have effective ess reduction.
The metering is helpful in optimizing the De-Esser’s settings. The orange meter shows the amount of reduction. The blue meter shows the input level.
How to Use the Shape Parameter
Ess sounds cover a fairly small range of high frequencies. The Narrow Shape is best for this application because it compresses a narrow band. Frequencies above and below that band remain untouched.
In addition to ess sounds, the De-Esser can also reduce “shhh” sounds (e.g., like the shh sound in “action” or “compression”). Shh sounds cover a wider range of frequencies, and often require a lower Frequency setting. For these sounds, the Wide Shape splits the audio into high and low bands, and processes the entire high band.
You may need to do two passes, one with a Wide Shape for shh sounds, and one with a Narrow Shape for ess sounds. Be conservative with the settings for the two passes, because the changes will reinforce each other.
How to Use the Range Parameter
This parameter is the De-Esser’s unsung hero. Setting Range to Full allows the full amount of reduction dialed in by S-Reduction. Gentle Range restricts reduction to ‑6 dB.
The Gentle setting is useful for more than just guaranteeing a subtler effect. With the Gentle Range enabled, you can dial in huge amounts of S-Reduction. This allows processing as much of the ess or shh sound as possible, not just the initial transient. However, limiting the amount of reduction to -6 dB prevents the amount of reduction from being objectionable.
Final Note
The De-Esser is not just for singing, but also podcasts, voiceovers, and narration. It can even reduce harshness with amp sims, as described in De-Esser Meets Amp Sims. And it can probably do other things that are yet to be discovered!
In the 60s, flanging was an electro-mechanical process that involved two turntables or two tape recorders. Since then, flanging has evolved into a digitally driven effect, with a variety of cool bells and whistles. Paradoxically, though, many of today’s digital flangers can’t reproduce the period-correct sound of 60s flanging.
Five years ago, I wrote a blog post about a vintage flanger FX Chain that took advantage of Studio One Pro’s Splitter and Extended FX Chains features. Although this week’s tip takes a little more effort to set up than just loading an FX Chain, and doesn’t have Macro Controls, it gives the same authentic flanging sound for Studio One Artist—check out the audio example.
The 60s flanging sound has three important qualities:
The Setup
Fig. 1 shows the track setup for the flanging effect.
1. The track you want to flange feeds two FX buses via pre-fader Sends. Turn the track’s channel fader down all the way. The Sends must have the same level (e.g., -6.0 dB).
2. Insert an Analog Delay in each FX Channel. Use the settings in fig. 2 for both of them.
3. Insert a Mixtool after the Analog Delay in the Variable Delay FX Channel. Turn on the Mixtool’s Invert Left and Invert Right buttons.
4. The channel faders settings for the Fixed Delay and Variable Delay channels need to be identical, and track each other. It’s a good idea to Group them.
How It Works
The Fixed Delay channel has a 5 ms delay. This is the “dry” channel. The Variable Delay channel’s Mixtool flips the FX Channel’s phase. The Analog Delay in the Variable Delay FX Channel can be longer or shorter than 5 ms. So, when the audio passes through 5 ms of delay, there’s the cancellation effect of through-zero flanging. By delaying the “dry” path, we’ve effectively allowed the Variable Delay to go into the future…at least as far as the dry audio path is concerned.
Create the flanging effect with the Analog Delay controls in the Variable Delay channel:
Customization
No matter how carefully you set up a mic’s pop filter, some pops are bound to get through the filter with vocalists who sing close to the mic. But you don’t need to redo or punch the vocal—let’s explore several ways to fix these problems in the mix.
Pro EQ3 dynamic EQ. Dynamic EQ is a fast, simple way to reduce pops (fig. 1):
1. Enable the low-frequency (LF) stage. Click the D button to reveal the dynamic EQ parameters.
2. Choose a 12 or 24 dB/octave low-frequency shelf.
3. Start with a substantial negative Range (e.g., ‑12 to ‑20).
4. Set Thrs. (threshold) to 0.00.
5. Loop the vocal section with the plosives. Adjust the low-shelf frequency so that the low-frequency attenuation affects only the plosives, not the voice’s usual warmth.
This technique also allows for a cool trick. Increase the Q slightly. When the plosives are being reduced, the increased Q adds a slight boost just above the shelf’s corner frequency. This gives some extra warmth to compensate for the lows being reduced.
Although this technique is simple to set up and works on an entire track, it may not be effective enough with super-nasty pops. The following methods are labor-intensive but can annihilate pretty much any plosive.
Split the Event before and after the pop, then insert Low Cut EQ as an Event FX. Roll off the low frequencies with the Event FX EQ (fig. 2). The only drawback is that if there are clicks when transitioning in or out of the isolated event, you’ll need to add crossfades or fades.
Split just before the pop, then fade in over the beginning (fig. 3). This reduces the level of the pop’s most prominent part—the beginning. For the best results, you need to find just the right split point prior to the pop, and carefully edit the fade shape.
For a solution that fixes plosives and sibilant (“ess”) sounds, check out The Vocal Repair Kit blog post. It’s also a tip in The Huge Book of Studio One Tips & Tricks v1.4 (see page 174).
This week, I wanted to give y’all a little gift: 15 “analog cab” IRs that provide alternate User Cabinet sounds for Ampire. Just hit the download link at the bottom, and unzip.
If you’re not familiar with the concept of an analog cab, it’s about using EQ (not the usual digital convolution techniques) to model a miked cab’s response curve. This gives a different sonic character compared to digitally-generated cabs. (For more information, see Create Ampire Cabs with Pro EQ2.) An analogy would be that convolution creates a photograph of a cab, while analog techniques create a painting.
The 15 impulse responses (IRs) in the Ampire Analog Cab IRs folder were made by sending a one-sample impulse through EQ, and rendering the result. This process creates the WAV file you can then load into Ampire’s User Cabinet. The IRs include the following cab configurations: 1×8, 1×10, (4) 1×12, (3) 2×12, 4×10, and (5) 4×12.
How to Use Analog Cabs
In any event, whether you go for individual impulses, layering, or creating stereo impulses, I think you’ll find that “analog” cab IRs extend Ampire’s sonic palette even further. And if you have any questions, or feedback on using analog cabs, feel free to take advantage of the Comments section!
Download the Ampire Analog Cab IRs.zip file below: