Basic Digital Recording 9780857121806


270 119 27MB

English Pages [135] Year 2009

Report DMCA / Copyright

DOWNLOAD PDF FILE

Recommend Papers

Basic Digital Recording
 9780857121806

  • 0 0 0
  • Like this paper and download? You can publish your own PDF file online for free in a few minutes! Sign Up
File loading please wait...
Citation preview

basic Digital Recording Paul White

Copyright © 2009 SMT This edition © 2009 SMT (A Division of Music Sales Limited, 14-15 Berners Street, London, W1T 3LJ) ISBN: 978-0-85712-180-6

The Author hereby asserts his/her right to be identified as the author of this work in accordance with Sections 77 to 78 of the Copyright, Designs and Patents Act 1988. All rights reserved. No part of this book may be reproduced in any form or by any electronic or mechanical means, including information storage or retrieval systems, without permission in writing from the publisher, expect

by a reviewer who may quote brief passages. Every effort has been made to trace the copyright holders of the photographs in this book, but one or two were unreachable. We would be grateful if the photographers concerned would contact us. A catalogue record for this book is available from the British Library. For all your musical needs including instruments, sheet music and accessories, visit www.musicroom.com

For on-demand sheet music straight to your home printer, visit www.sheetmusicdirect.com

basic Digital Recording

Paul White

smt

Also by Paul White and available from Sanctuary Publishing: Creative Recording 1 - Effects And Processors (second edition) Creative Recording 2 - Microphones, Acoustics, Soundproofing And Monitoring (second edition)

Home Recording Made Easy (second edition) EMERGENCY! First Aid For Home Recording MIDI For The Technophobe (second edition)

Recording And Production Techniques Also by Paul White in this series: basic basic basic basic basic

Effects & Processors Home Studio Design Live Sound Mastering Microphones

basic MIDI

basic basic basic basic basic basic

Mixers Mixing Techniques Multitracking Sampling VST Effects VST Instruments

CONTENTS

Information Page Introduction

1

Digital Basics Sampling Theory Sample Rates Sample Rate Converters

Digital Connections Clock Sync Word Clock Digital Effects Digital Reverberation Reverb Parameters Gated Reverb Delay And Echo Modulated Delay Chorus And Adt Flanging Phasing Vibrato

Pitch Shifting Autopanners Amp/Speaker Simulators Filters MIDI Control

Patching Effects Processors

Compressors/Limiters Gates Equalisers

Exciters

Recording Systems Audio Quality Input/Output

Random Access Non-Destructive Editing Editing Features Hard-Disk Limitations

What Can A Computer Do? Rewire

Software Samplers Hard-Disk Editors Practicalities Recording Hardware Choosing A Computer Hardware Recorders Digital Tape Tapeless Workstations Digital Mixers Stereo Recorders MiniDisc SCMS CD-R

The Soundcard Studio Macintosh Users Connections Recording Pre-Amps And There’s MIDI Evolution Software Synthesis Latency And Monitoring Input Monitoring Punching In Choosing A Soundcard Soundcard Drivers BeOS No Games!

Music Software The Sequencer The MIDI Composer

MIDI And Sequencing Audio Sequencers Sequencer Setup Metronomes

Tracks And Channels Cut, Copy And Paste Playback Editing Quantising MIDI Drums

Types Of Sequencer MIDI Data Storage

Computer Complexity

The MIDI Interface The User Interface Extra MIDI Ports

Common Problems MIDI Timing Sequencers With Audio

Software Plug-Ins Third-Party Plug-Ins Native Plug-Ins

Multiple Plug-Ins Types Of Plug-In Plug-In Automation Software Protection Computing Power Stereo Editors Software Synths Samplers

5

Music Computers Recording Levels Weak Links Drive Maintenance Effects

Other Plug-Ins Guitars And Computers Fan And Drive Noise Waveform Editing Computer Power Time And Pitch Manipulation Keep It Clean!

Computer Timing Lightening The Load Audio Timing

Audio And Clocks Sync

Solutions

6

Creating An Album Mastering Mastering EQ Gentle Compression

De-Noising Track Editing Making A CD The Internet

Appendix: Common Cable Connections

Glossary

INTRODUCTION

Today the majority of recording systems are digital, but unlike in the era of analogue tape there’s no longer a single accepted approach. Instead we are faced with a variety of technologies, some using tape, some using computer hard disk drives and some using MiniDiscs. Most of the cheaper systems are used with the computer desktop, on which digital recording is often combined with MIDI sequencing, but there are

also

stand-alone

hard-disk

recorders

and

digital

tape

recorders available. The type you choose will depend largely on the type of recording you wish to make and on whether or not you need to edit your recordings afterwards. The aim of this book is to explain the theory of digital recording and to discuss the various types of recording system currently on the market so that you’ll be in a better position to decide which system is right for you. There are also numerous tips on how to keep computer-based systems working at maximum efficiency, along with some guidelines on recording and mixing.

1 DIGITAL BASICS

Before working with a digital recording system, it’s important to learn a little background information about how they work and how they differ from their analogue counterparts. All digital audio equipment designed to accept analogue signals, such as the outputs from microphones, must convert the analogue input into a digital format before it can process it or record it. An analogue signal is one in which the voltage varies in proportion to the original event; in the case of sound, the signal voltage changes in proportion to fluctuations in air pressure. A rapidly vibrating string, for example, will create rapid fluctuations in air pressure which a microphone can convert into variations in voltage. A digital system, on the other hand, works with binary numbers - ones

and zeros - which are represented in the circuit by the presence or absence of a nominally fixed voltage. Converting an analogue signal into digital information involves measuring the analogue voltage at regular intervals and then turning these measurements into a series of binary numbers. If there are enough instantaneous measurements per second the original sound can be recreated. So how many measurements

range

or samples

of human

hearing?

are necessary

This

is where

to capture

sampling

the full

theory

comes in.

Sampling Theory Measuring and digitising minute sections of an input signal is known as sampling. Figure 1.1 illustrates what happens when

a signal is sampled. First it is cut into slices, rather like a loaf of sliced bread, and the height of each slice is then measured. As shown in the illustration, the slices, or samples, have flat

tops (corresponding to the voltage measured during the sampling time), so they don’t accurately follow the curve of the waveform. From this you might deduce that the representation of the original sound would be more accurate with thinner slices, and up to a point, this is indeed the case. Sampling theory states that, if the output is to be reconstructed accurately, samples must be taken at a minimum of twice the frequency of the highest harmonic likely to be encountered. If the sampling frequency is less than twice the highest frequency, additional frequencies will be introduced based on the sum and difference between the sampling frequency and the audio frequency. These frequencies won’t have been originally present and will sound musically unpleasant. This effect is known as aliasing, and to prevent it, any frequencies in the original signal that are above half the sampling frequency must be filtered out. However, because no filters are perfect, the sampling frequency will be a little higher than twice the maximum audio frequency, which is why, on an audio bandwidth of 20kHz, the minimum sampling frequency is 44.1kHz. If you're thinking that 44.1kHz is an odd kind of number (why not 45kHz or 50kHz?), you’re right. As is so often the case, the reasons for using this frequency are based on convention.

The signal voltage is sampled at regular intervals so that the original signal can be reconstructed by

in effect joining the dots. The sample levels are converted to binary numbers for digital sterage

and manipulation ;

hE

= |

Original waveform

>

A!



,

Voltage

Time a

-ve

Figure 1.1: Sampling a waveform

Audio quality is also affected by the accuracy with which samples are measured, and if more digital bits are used to represent each sample then the accuracy of these measurements

increases.

CDs

and

DAT

tapes

use

16-bit

sampling, though many digital multitrack recorders signal processors now use 20- or 24-bit conversion.

and

Digital numbers are separated by discrete steps - there are no halves or thirds of a bit - and the number of steps used depends on the resolution of the analogue-to-digital converter used. Eight bits will give you only 28 steps, which works out at 256. That means that your loudest signal could have 256 steps, but quieter sounds will have considerably fewer. This value provides a rather poor level of resolution and

causes

Nowadays,

what’s

known

eight-bit sound

as _ quantisation

is rarely used,

except

distortion.

in some

undemanding applications. Quantisation distortion sounds like noise, and is in fact very

similar, except that, unlike most other noise, it disappears in the absence of a signal. Using more bits results in a vast improvement in resolution, and most current digital processors use 16-bit (like compact discs), 20-bit or even 24-

bit sampling. A 24-bit system samples with more than 16 million steps, which makes the 256 steps used by eight-bit sampling look a little inadequate! Each bit used in a conventional sampling yields a maximum dynamic range of 6dB, so at best an eight-bit system can only provide a dynamic range of 48dB, which is about as noisy as a

cheap

cassette

recorder.

16

bits,

on

the

other

hand,

provides a maximum dynamic range of 96dB, which for most audio purposes is more than adequate, and even though 20and 24-bit systems aren’t yet perfected, they can still provide practical dynamic ranges in excess of 120dB.

Sample Rates Although the minimum sample rate for serious audio work is 44.1kHz, less critical work is sometimes carried out at 32kHz (which reduces the audio bandwidth

to under

15kHz).

Most

broadcasters prefer to work with a sample rate of 48kHz. Most DAT machines and soundcards support sample rates of both 44.1kHz and 48kHz, but it’s important to note that everything else in the system must be set to run at the same sample rate, or the final audio recording will not play back at the same speed as that at which you recorded it. If you’re planning on making CDs then you should always work at 44.1kHz, while if you’re producing material for broadcasting you should stick to 48kHz. Recent systems have been introduced that run at twice the

current standard sampling rates, so now 88.2kHz and 96kHz can be added to the list. In theory these systems produce a slightly better sound quality, but in practice few people can detect a difference. Even so, it’s important to be aware of them, especially as the 24-bit 96kHz audio used by DVD may become the standard hi-fi format.

Sample Rate Converters Sample rate conversion is necessary in those situations when you want to work at one sample rate but some of the source material is configured to another. It may be accomplished either with software or with an external hardware box, which

will take in a digital signal at one sample rate and output it at another in real time. To recap, then: systems with higher sampling frequencies can handle higher input frequencies, which in turn means a better frequency response. The only trade-off is that, because you end up recording a higher number of samples per second, more memory is needed to store the extra sound data. The same is true when recording more bits per sample. (The number of bits per sample is sometimes known as bit depth.)

Fortunately,

RAM

(Random

Access

Memory)

and

computer hard drives are orders of magnitude cheaper now than they were a few years ago, so data storage is no longer the major cost concern it once was.

Digital Connections Unlike analogue systems, on which every signal needs to be sent down a separate cable, digital systems allow two or more signals to be sent along the same cable while remaining completely separate. It’s not important to know the theory behind this apparent magic at this stage, but it’s important to know what is meant by the terms ‘clocking’ and ‘sync’.

Let’s say that you have a computer studio setup with a digital input and you want to transfer some music from a DAT (Digital Audio Tape) recorder into your computer, so the material enters the digital domain. The digital output from your

DAT

machine

(probably

an

S/PDIF)

phono

connector

must first be connected to the digital input of your computer

interface or soundcard. S/PDIF is a standard connection protocol developed by Sony and Philips to be used with consumer digital equipment, and on some it may be provided as an optical port, as an alternative - or in addition - to a phono co-ax connector. There are commercial-format adaptors available which convert co-axial S/PDIF to optical, and vice versa. This is where clocking comes in. When the DAT machine is played, a highly accurate crystal clock controls the rate at which the samples of data are played back to make sure that the material is the same length and at the same pitch as the original. To feed these little slices of data into a computer, the computer’s sampling clock must be running at exactly the same

rate

as

that

of the

DAT

machine,

but

if the

DAT

machine and the computer are allowed to run independently, their clocks will always run at slightly different speeds, even if they’re both set to the same frequency. This produces clicks and crackles in the audio as the two clocks drift in and out of sync.

Clock Sync Fortunately, the digital data carried along an S/PDIF cable also carries the clock signal from the first machine, and this is used to sync the clock in the receiving device. However, the receiving device usually has to be set to Digital Sync mode for this to happen (see Figure 1.2). DAT machines without a word clock input facility switch to Digital Sync mode automatically if they are switched to digital input,

which is sensible because there are no normal circumstances in which you’d want to accept a digital input while the receiving DAT machine was running from its own internal clock

(Internal

Sync

mode).

Nevertheless,

some

computer

systems require you to select Digital Sync mode manually. The rule is that the source device runs from its own clock (Internal

Sync

mode)

and

the

receiving

device

is

set

to

Digital Sync mode. A number of devices can be cascaded in this way, but stringing more than four digital devices together in a chain may cause problems if the clock signal is corrupted.

S/POIF digital output Ff

Computer S/P OIF

digital input

Computer soundcard 36 to External Digital Syn moda Figure 1.2: Digital connections (S/PDIF with sync)

Note, however, that more professional equipment is based on the AES/EBU digital interface, which can usually be identified by the presence of XLR connectors instead of phonos. In any event, the sockets are normally marked

clearly on the rear panels of the equipment. Both S/PDIF and AES/EBU systems carry a stereo signal and a clock, but they operate at slightly different levels and so are not strictly speaking compatible (even though it’s sometimes possible to use

an

adaptor

lead

to

connect

the

two).

AES/EBU

is

a

balanced system and can use conventional mic cables over short

distances,

while

S/PDIF

is

an

unbalanced

70-ohm

system that requires proper digital 70-ohm coaxial cable to work properly. You may be able to use audio phono leads over very short distances, but this isn’t recommended because, although everything may seem to work fine, it may produce the occasional click in the audio stream.

Word Clock Most professional systems have a word clock facility, which simply means that the sample clock is fed via separate sockets rather than being forced to rely on the clock embedded in the digital audio signal. In complex setups, this can be more accurate and less prone to clock corruption or problems involving jittery timing. Word clock usually travels via unbalanced cables capped with bayonet-fitting BNC connectors, and like other digital cabling it should be clearly labelled on the rear panel of the equipment. As with other digital clocking systems, there can only be one master - all the other pieces of equipment act as slaves, locking to the master clock. In large studios the word clock may come from a separate, high-accuracy clock generator with multiple outputs, synchronising everything in the studio, but in a project studio system it’s more likely to originate from the digital console, if there is one. It’s also common for systems to include

master

a mixture,

word

clock

where

while

some

others

devices

(usually

are locked

without

clock input) are syne’ed via S/PDIF or AES/EBU.

to a

a word

Digital Effects In the modern studio, digital electronics are used to record,

process and mix sound. Effects and signal processors play a large role in contemporary music production, and now that digital circuitry can do virtually everything that analogue circuitry can, most hardware processing boxes have software equivalents that are available as plug-ins for the main audio software. Most studio effects are based on delay (echo, reverb, flanging, chorus and so on) or on pitch shifting. Processors, however, such as equalisers, compressors and gates, are more concerned with the dynamics of a signal. An overview of these basic effects follows.

Digital Reverberation Reverberation occurs naturally when sound is reflected and re-reflected from surfaces within a large room, and an electronic reverb unit mimics this effect by electronically generating thousands of reflections. Reverb, like all modern effects, can be used to create the impression of sounds being recorded in a room, but it may also be used to create new effects that have no obvious counterpart in nature. Sounds produced in different types of rooms and with different materials acquire different qualities, which is why modern reverberation units provide a number of room types and user-adjustable parameters. Most applications require a fairly short reverb time of between one and three seconds, although digital reverbs can also emulate caverns with decay times of ten seconds or even more. Plate settings are popular for general use, especially on vocals and drums. The term refers to the mechanical reverb plate that was used before digital reverb units were invented. In real life, each ear picks up a different pattern of reverb reflections, providing the brain with stereo information. Digital reverb units process a mono input to produce two

different sets of synthetic reflections in order to create a convincing stereo effect. Figure 1.3 shows the pattern of decaying reflections created by a typical digital reverberation unit.

Good digital reverbs can produce excellent approximations of sounds being generated in real acoustic spaces, but they also have the capacity to create reverberant characteristics that simply couldn’t occur in nature. Digital systems also allow us to produce reverb that is longer and brighter than anything normally encountered in nature. Amplitude

Impulse such as a

handclap

ra

Time

Initial Early

Later dense

delay

reflections

reflections

Figure 1.3: Reverb reflections

Reverb Parameters The main reverb parameters available for user control are: pre-delay time, pattern and level of early reflection, overall decay time and high-frequency damping. Pre-delay simply determines the time between the original sound and the first reflection, and can be anything from virtually instantaneous to half a second or more. Early reflection patterns are not usually variable. Normally the user can only select from a handful of stored patterns or

presets simulating rooms of various sizes, including halls, chambers, plates or small-room ambience, although in some

cases the level and spacing of these reflections can be altered. The impression of room size is increased if the reflections are spaced further apart. Overall decay time simply determines the time it takes for the reverb to die away. Longer reverb times are suggestive of large environments with reflective surfaces, while shorter times may be used to simulate the acoustics of an average small room. Most reverb units can produce an impressively long decay time, but the most accurate way of judging the quality of a reverberator is to hear how convincingly it emulates small-room ambience. The use of high-frequency damping means that the highfrequency decay time can be shorter than the overall decay time, and can therefore be used to simulate the absorbency characteristics of real rooms, simulating the way in which both surface and air absorption affect sound. Some units also have independent control over low-frequency damping, and can simulate environments that reflect mainly high-frequency sounds. By selecting the right pattern for the environment you wish to simulate and then adjusting other parameters, the available effects can vary from a barely reverberant room to a huge hall where the reverb decay thunders on for several tens of seconds. In practice, because long reverbs tend to muddy a mix, most of the useful treatments have a decay time below two seconds, although there are occasions on which a long reverb can be effective. Some models incorporate a control which determines the room size of the reverb by simultaneously adjusting several

parameters occurring in to have, as reprogram a

in order to provide the impression of sounds a larger or smaller space. This is a useful control it saves time and means that you don’t have to number of parameters manually.

Gated Reverb As well as the simulation of natural reverberation, two more

effects have become standard issue in the modern reverb unit: gated reverb and reverse reverb.

digital

Gated reverb was first created by using the (often heavily compressed) ambience of a live room and a noise gate to produce an abrupt cutoff point rather than a smooth decay. Most reverb units provide an electronic emulation of this effect, using a burst of closely spaced reflections that stop abruptly after around half a second. Reverse reverb doesn’t involve playing anything backwards, as the name might suggest, but is instead achieved by applying a reverse envelope to a group of reflections so that, instead of decaying, they increase in level after the original sound before cutting off. Like gated reverb, the most important parameter is the time it takes for the reverb to build up and cut off.

Delay And Echo A delayed sound is produced by delaying an audio signal in order to produce one or more distinct echoes. This is the electronic equivalent of a tape echo unit. A feedback or regeneration control adjusts the amount of output signal fed back into the input so that repeating echoes can be produced. To set up a single delay, the modulation depth, rate and feedback controls should be at their minimum settings, and the decay time may then be set as desired. To convert this single repeat into a true repeating echo is simply a matter of

advancing the feedback control so that whatever appears at the output is then fed back to the input, returning and forming a delay loop. The feedback control determines the time it takes for the echoes to die away.

Modern multi-effects units often include multitapped delay programs, which generate a number of echoes set at different delay times. With these, it may also be possible to pan the individual delays from left to right in the stereo field to create interesting spatial effects.

Modulated Delay Originally

conceived

machines,

delay

soon

acquired

as the

devices

successors

(and

modulation

their

controls

of tape

software

that

loop

echo

counterparts)

allowed

them

to

produce a wide range of effects, from echo and doubling to chorus, flanging, ADT, vibrato and phasing. Indeed, many of the standard

studio

effects

are variations

on the theme

of

digital delay, where the delay time is modulated by an LFO (Low-Frequency Oscillator) to produce effects such as chorus,

flanging, phasing and vibrato.

Chorus And ADT Chorus uses a short delay of up to about 7Oms to create a slight doubling effect, and the delay time is then modulated to produce a slight wavering in pitch. An equal mix of the delayed and unprocessed sound is used to produce an interesting effect, which sounds rather like two instruments playing the same part but with slight differences in timing and tuning. ADT

(Automatic

Double

Tracking)

is similar

to chorus

but

uses a delay time in excess of 70ms to create a more pronounced doubling or ‘slapback’ effect, and the depth of modulation is shallower. ADT is often used to process vocals

in order to obtain a thicker sound, making it sound like a singer has performed the same part twice, on different tape tracks. In this case, the amount

of modulation

should be so

slight that it’s only just noticeable.

Flanging Flanging is essentially similar to phasing, although it may use slightly longer delay times - up to 50ms, for instance - and the feedback control is advanced to produce a dramatic, swirling effect. In general terms, it’s possible to get away with using more depth by using a slower modulation rate. The sound becomes more ‘whooshy’ if more feedback is applied.

Phasing Phasing uses shorter delay times than flanging, and little or no feedback, to produce a moving comb filter which sounds similar to mild flanging. Because the effect is more subtle, however, it can be used more extensively. A lot of music recorded in the ‘70s features phasing on the lead guitar part.

Vibrato Vibrato is a modulation of pitch similar to that produced manually by a guitar or violin, and is created by using only the delayed sound and none of the original. The delay is kept to just a few milliseconds so that the timing of the performance isn’t significantly affected, and the depth of modulation determines the depth of vibrato.

Pitch Shifting Pitch shifters can change the pitch of the original signal without changing the speed of the sound, and they usually have a maximum range of at least an octave above and below the original sound. Most budget pitch shifters impart a

strange timbre, which is caused by the regular modulation of the looping processes, but this side-effect can be disguised by mixing it with the original sound. There’s also a slight delay due to the looping process, but this can be as short as just a few milliseconds. Smaller pitch shifts sound very similar to chorus, albeit without the chorus effect’s regular modulation. Such detuning treatments - combined with a short delay - are also often used to double or thicken vocals.

Autopanners Autopanners are devices that automatically pan mono signals from left to right in the mix, usually under the control of a LFO or an external trigger. Used in time with the tempo of a track, panning can be applied quite subtly.

Amp/Speaker Simulators While keyboards usually work best when played through a sound system with a flat frequency response and minimal distortion,

means

guitar

and

bass

that their frequency

instrument

rather

than

amplifiers

response left

flat.

are

voiced,

is shaped

which

to suit the

Furthermore,

the

loudspeakers and enclosures used in guitar and _ bass amplification usually have a very limited frequency response, which enables them to filter out the rougher-sounding components of amplifier distortion. If you were to DI (Direct Inject) a distorted guitar without EQ, the result would be a much thinner and more raspy version of what you would hear from an amplifier. To make DI’ing the guitar a more practical proposition, the amp/speaker simulator was devised. This device operates by using a filter circuit to mimic the amplifier and loudspeaker voicing of a typical guitar amplifier. Many multi-effects units now include amp/speaker simulators as well as overdrive, which enables the user to create a fully produced guitar sound with only

one unit. The output may then be recorded without further processing.

Filters Effects units often include filters, not only in the form of parametric and graphic EQ but also as emulations of the swept-resonant filters used in synthesisers. The filter used in a typical analogue synthesiser is closely related to the parametric equaliser - the main difference is that the frequency of the filter can be controlled electronically rather left to the user. For example, an LFO could be used to sweep the frequency up and down, or an envelope could be generated to provide a filter sweep. Also, the filter can often be set to a higher Q (resonance)

value - some

can be set so

high that they self-resonate and turn into an oscillator. Some of the more sophisticated multi-effects processors include not only resonant synth-type filters but also a variety of possible control sources, including envelopes derived from the input signal level, MIDI-triggered envelopes, LFOs and so on.

MIDI Control MIDI-plus-audio

sequencers

can

use

MIDI

information

to

automate certain effects parameters (as long as the machine is equipped with a MIDI control facility), and it’s also possible for many plug-in effects to be automated. All but the very cheapest digital effects units are now programmable, offering MIDI control over patch selection as well as some form of dynamic MIDI control over effect parameters. The ability to program effects enables the user to set up several effects of each type, and with chorus, flanging,

ADT,

vibrato and other DDL-based

effects, these programs,

once created, can be used in a number

of different contexts

with little or no further modification. If there are more than 128 presets and user patches, these are normally organised

into banks so that they can be accessed using MIDI Change messages.

Bank

Patching Effects Effects

such

as reverb

are normally

fed from

a post-fade

(effects) send control, which means that, if the channel fader

is moved, the reverb level will increase and decrease along with the original sound. This is obviously the way you’d normally want to work, although it’s possible to drive a reverb unit from a pre-fade send if you want to fade out the dry (uneffected) signal while still leaving the reverb audible. This procedure is useful in producing unusual or special effects. The aux send system has the advantage that it allows the same effects device to be shared among all of the mixer channels while still allowing the user to have different amounts of effect on each channel. For example, you might be able to use the same reverb setting on both drums and vocals, but you may want to use more on the vocals than on the drums.

Processors Examples of processors include exciters, equalisers, compressors, limiters, gates, expanders and autopanners. All

of these devices change either the level or the tonal quality of the signal passing through them, and you wouldn’t normally add a dry signal to the output of one of these units. Because aux sends are used to add the output from an effect to the dry signal, it isn’t common practice to connect a processor via an aux send; instead they are connected by an insert point - usually into a mixer channel or group. Most of these processors were originally designed to be analogue, and so will be described as such where appropriate. However, they are all available in digital versions in both hardware and

software form.

Compressors/Limiters Compressors are used to even out excessive peaks in signal level that occur in vocal or instrumental performances, and

they do this by changing the gain of the signal path depending on the level of the signal passing through. Just as an engineer might pull down a fader if the level gets too high, a compressor will lower the level if it exceeds a threshold set by the user, so that signals falling below the threshold level remain unchanged. The degree of gain reduction applied is set by the compression ratio control setting, as shown in the graph in Figure 1.4. Output

Z .

level (dB)

17 rat rate

if a

#

Gain reduction takes Wace

above

the

threshold

Ps

a

_—

el

Threshold

Pes

ft

’ a

fo

i

ratio

_. #1 ratio —

af Vs Fr

2:1

§.1 rate

Infinity:1 ratio {hard limiing)

The output signal is E unchanged below

the threshold

f

wf ft fo

Lf

input level {a5} Figure 1.4: Compressor graph

When the ratio is very high, the compressor’s maximum output is maintained at the threshold level and is effectively prevented from going beyond it. This imposes a limit on the maximum signal level, regardless of the input, and a compressor used with a high ratio setting is often described

as a limiter, hence the term compressor/limiter. As a rule, high compression ratios are used in conjunction with high threshold settings, so that only the signal peaks are affected. Lower ratios may be used with lower threshold settings to provide gain levelling over a greater dynamic range. The compressor’s attack control determines how quickly the circuitry responds to a signal once it has exceeded the threshold,

while

the release

control

determines

the time

it

takes for the gain to return to normal once the input has dropped back below the threshold. Most two-channel compressors have a Stereo Link switch so that stereo signals can be processed. In Stereo Link mode both channels are compressed by the same amount, thus preventing the image from appearing to move to one side or the other when a loud sound appears on only one side of the mix.

Limiting is used in those circumstances

when

it would

be

undesirable for a signal to exceed a specific level-to prevent

clipping on a digital recorder, for example.

Gates Whereas compressors control the levels of signals that exceed a threshold, gates control the levels of signals which fall below a threshold. The aim is to silence the signal during pauses, when background noise isn’t masked by a signal. If the gate’s threshold is set at a level just above the background noise, the gate will operate whenever there is a pause in the signal. Gates are used to mute spill from other instruments as well as

to tackle straightforward noise or hiss, and they are used regularly in the miking of drum kits-for example, in preventing the bass drum mic from picking up the other drums in the kit. They are also often used with electric

guitars, combating the high levels of noise and hum generated by guitar amplifiers when the guitarist is not actually playing. Vocals may also be gated to hide the sounds of rustling clothing and breathing. It must be understood that gates only remove noise when there is a pause in the wanted signal - they can’t help with noise that’s audible over the top of the programme material. Figure 1.5 describes a gate’s reaction when the signal falls below the threshold level. Expanders are very similar to gates, except that they close down gently, more like a compressor in reverse. Gaate remalns open until tha hold time has elapsed

|



signal ,

level

Gate fully opan

Hold time

Original

Gales clesés at

rate sel by

audio

release control

|

anvéelone

ba!

j

Threshald—+-—-

\, _

Signal after gating =

Gate starts ta open

fe

rime e

Signal falls

below threshold

Wantad signal

last whan gata is

closed

Figure 1.5: Operation of a gate

Equalisers Equalisers are essentially tone controls. Most mixers are equipped with built-in EQ facilities, but it’s often useful to have a high-quality external equaliser (or its software counterpart) that can be patched in via an insert point for those

occasions

when

more

precise

control

is

required.

Parametric equalisers are the most versatile of the lot, but they also have the disadvantage that they take a long time to

set up properly. Another book in this series, basic Effects & Processors, provides and in-depth analysis and comparison of these devices and their applications.

Exciters Also

known

as

enhancers,

exciters

are

devices

that

add

synthesised high-frequency harmonics to a signal in order to make it sound brighter and more intimate. The process, which was developed and patented by the company Aphex, is different to the process of equalisation, which can only redistribute those harmonics that already exist. Exciters are used to push sounds to the front of a mix, or to create an impression of clarity and space in crowded mixes. They can be used on both entire mixes and individual tracks, and may also be used to brighten stereo mixes which are destined to become masters, for the mass-copying of cassettes, in order to compensate for the loss of high-frequency sounds that cassette duplication often entails. Enhancers working on slightly different principles are available from _ other manufacturers.

2 RECORDING SYSTEMS

Only a few short years ago tapeless recording was too expensive to contemplate, but now the cost of the hardware involved has come in down

in price so much that it is now

the least costly of all of the options available. However, harddisk recording is accompanied with its own set of choices and challenges - for example, which computer should you choose? And if you prefer to use hardware, which company provides the best equipment for your requirements? Should you decide to use a computer recording system, it’s possible to record audio onto a Macintosh with nothing more than the appropriate software (audio inputs and outputs are built into all current models),

or you

can record

onto a PC

fitted with a suitable soundcard and software. Alternatively, you can use a computer with additional external hardware, such as Digidesign’s Pro Tools III, thus relieving the computer of most of the burden of processing and mixing audio. If you prefer to use hardware, however, there are standalone, hard-disk-based digital workstations available which offer upwards of eight tracks, and many of these machines cost little more than their obsolete analogue-cassette-based counterparts. Prices have fallen significantly because, unlike analogue recording, the hardware used in tapeless systems is based on the same components that are used elsewhere in the computer industry, which means that the economics of scale are, for once, on the side of the consumer.

Audio Quality

16-bit hard-disk recording can provide the same quality of audio reproduction as a CD, but it’s not always wise to take that level of performance for granted because the quality of the fidelity is also limited by the quality of the analogue-todigital converters, and indeed to how they are used. For example, high-quality converters in an external rack box are likely to sound every bit as good as a CD, while budget converters housed in a computer - either as part of the basic hardware or on a cheap soundcard - are more likely to pick up some noise from the computer’s own electronic components. They may also use budget converter chips, which offer relatively poor resolution and cause increased noise and distortion. Most professional hardware can now work at 16- or 24-bit resolution, and support for the 96kHz sampling rate is becoming commonplace, but a well-designed 16-bit system that can run at either 44.1kHz or 48kHz is more than adequate for most applications. Ultimately, the CD format is limited to 16-bit, 44.1kHz resolution - and you know how professional a well-recorded CD can sound!

Input/Output Input/output (or I/O for short) describes both the analogue and digital inputs and outputs of an audio system. The amount of I/O you need depends on your working methods if you only ever record one part at a time and intend to handle everything with the computer then you may need nothing more than a stereo in/stereo out card. On the other hand, if you plan to use an external mixer you may need a system with eight or more discrete analogue outputs, and if you're looking to record several parts at the same time you'll need a card with multiple analogue inputs. Users with digital mixers or digital tape machines may benefit from a system that offers A-DAT lightpipe compatibility, which allows eight channels of digital audio to

be transferred along a single fibre-optic link. In any event, it’s usually wise to choose a card that has stereo digital ins and outs (usually S/PDIF) so that it’s possible to record information to and from formats such as DAT,

CD/CD-R

or

MiniDisc. As you’d expect, systems with more I/O capabilities are more expensive, so think carefully about your current and future needs before deciding on which system you're likely to need.

Random Access The advantage that hard-disk recording has over tape recording is that computer hard drives can reach any piece of data within a few milliseconds - you don’t need to wait for a tape to wind to your chosen location. This attribute is known as random access (unlike the linear medium used by tapebased systems). Modem AV drives are equipped with intelligent thermal recalibration systems which avoid interrupting when the disk is reading or writing data, making them a better choice for multitrack work. Virtually all modern hard drives are AVcompatible - you should be careful if you find an old drive in the cupboard that you think might be useful. As a rule, EIDE drives (Enhanced Integrated Drive Electronics), which are used inside most Macs and PCs currently on the market, are slower than SCSI drive (Small Computer System Interface pronounced ‘scuzzi’), although there are also. several variations of SCSI available, with SCSI II being logically faster than SCSI I. Future systems are likely to rely on the FireWire connection protocol, which is significantly faster than SCSI, in which case the ultimate limitation is that of the

speed of the drive mechanism itself. Even so, a cheap EIDE drive should let you play back at least eight tracks of 16-bit, 44.1kHz audio simultaneously.

Non-Destructive Editing Because random access allows you to call up any piece of data almost instantly, you can rearrange the sections of your recording and play them back in any order. Rearranging, copying and removing material without in any way affecting the original is known as nondestructive editing, and there’s a lot that can be done to a recording without the original data changing in any way. Even so, there are times when it may be desirable for a change to be permanent - for example, when silencing noise immediately before the start of a song, or when changing the actual level of the recorded audio. This type of edit is destructive, because the original data has been changed.

Editing Features The most effective editing tricks available for digital systems are based around the rearranging of material into a different order, although it’s also possible to normalise signal levels, bring peaks up to OdB, apply digital EQ, reverse sections of sound, create fade-ins and fade-outs, and a number

of other

functions which would have been difficult or impossible on a tape system, whether digital or analogue. Most systems are now capable of supporting software-based plug-in effects, such as reverb, delay, compression and stereo enhancement. However, before I get too enthusiastic, it’s only fair that I should warn you of the disadvantages of recording to disk.

Hard-Disk Limitations Over half an hour of eight-track audio can be recorded onto a cheap VHS video tape, but even with the falling cost of hard drives it would still cost you a lot more than the price of one of these if you wanted to leave the recording on the drive once you’d finished it. There are removable disk systems available which can store data on removable cartridges with

capacities of a couple of gigabytes or more, but they still cost significantly more than tape-based systems on a per-trackminute basis. With recording time, it’s a general rule that one

track-minute of 14-bit, 48kHz audio eats up around 5Mb of memory. Because hard drives aren’t yet cheap enough to use for permanent archiving (in a project studio environment, at least), some form of backup system is needed. CD-R (Recordable

Compact

Disc) is good for storing up to around

two track-hours of audio, and for those who need a greater capacity, writable DVD (Digital Versatile Disc) is becoming cheaper all the time. Many professionals still favour the Exobyte tape backup systems used by the computer industry. Backing up material onto digital tape is obviously quite slow, but the relative cheapness of the equipment often makes it worth the inconvenience. It’s important to make backups because, if a disk becomes corrupted after a computer crash, all data could be lost. If a section of tape becomes

damaged,

however,

the rest of the

reel or cassette will still be playable.

What Can A Computer Do? Because computer audio technology is evolving at such a rate, you may not be aware of all the things of which a desktop music system is capable. Computers were first drafted into musical circles to handle MIDI sequencing, and then multitrack audio was added. The main benefit of a

computer-based system is that digital audio can be integrated into MIDI sequencing packages; any recorded audio can be accessed and arranged in the same way as sequenced MIDI patterns are arranged. This means that you only have to pay for one computer to work on both, and you don’t have to keep changing your mental outlook as you switch from MIDI

to audio.

If you already work mainly with MIDI then adding a few tracks of hard-disk audio is a very practical and flexible way to upgrade, as it removes the need to synchronise a tape system

to

a

MIDI

sequencer.

Because

a

typical

MIDI

composition may only include a small amount of audio (vocals and guitar parts, for example), removable drives become more cost effective as a storage medium. My own preference,

however,

is

to

complete

an

album

on

a

conventional, fixed hard drive and then back up both the audio files and the MIDI data onto one or more CD-Rs. Software mixing, EQ, signal processing and effects have now been brought into the computer domain, with effects and processors available as plug-ins. Plug-ins are separate pieces of software that can be added to your main audio program to increase its functionality. However, unlike most add-on software, the new functions appear from within your main audio program. For example, the widely adopted VST plug-in format set up by Steinberg allows the plug-ins to be used from within the digital mixing section of the software, usually via insert points in the channels or groups. Plug-ins are generally purchased separately, from _ third-party manufacturers with different specialities, but programs such as Logic Audio and Cubase VST and SX come with some genuinely useful plug-ins to help you get started. Note that Digidesign’s Pro Tools system uses its own format of plug-in (TDM)

because

the plug-ins have to be written to

run on their DSP (Digital Signal Processing) cards, rather than the host computer’s processor. A few other systems exist that require a proprietary plug-in format, but VST has taken off as a standard for most audio/MIDI sequencers on both Mac and PC.

As the processing power of computers continues to increase, designers are never far behind in thinking up new, creative ways of using it. Software synthesisers have become popular in recent years, and some are surprisingly powerful, even though it’s difficult to obtain the right software drivers. Steinberg’s ASIO standard is widely supported, which means that the output from a synth can be sent via a regular soundcard. Software drivers are required whenever hardware needs to interface with software - it acts as a translator between the two. All soundcards need drivers, as do printers,

CD-R recorders and external MIDI interfaces.

ReWire To make it possible to use a synth at the same time as a sequencer, Steinberg (in conjunction with Propellerhead Software) have provided a solution in the form of ReWire,

a

system which allows these virtual synths to be routed inside any Mac or PC MIDI/audio sequencer that supports it. ReWire effectively allows a synth to be routed through the audio software’s mixer section just like any other audio track, and it can handle up to 64 channels of real-time audio at once.

Software Samplers There are currently a few software samplers on the market for both Mac and PC, and I envisage that software sampling and sample playback will be an area of significant growth over the next couple of years. We may even see sampling offered as a standard feature on MIDI sequencers before too long. When you consider that computers can already handle CD

burning

(writing

material

onto

CD-Rs)

quite

easily,

it

should be of no surprise that a well-thought-out computerbased desktop system can handle a complete music project, from recording to CD mastering, without the need to use

additional

hardware

other

than

mics,

pre-amps

and

a

monitoring system.

Hard-Disk Editors Hard-disk editors exist specifically to manipulate and rearrange stereo (or sometimes surround sound) material, and they are most commonly used to compile individual tracks in order to produce a production master tape for an album project, or to chop up and move the various sections of a song around in order to create a new arrangement. Such editing techniques are often used to create extended remixes or to shorten songs. Sophisticated crossfading algorithms are used to ensure that there are no audible glitches at the edit points - something that isn’t always possible when editing analogue tape with razor blades! Hard-disk editing is also frequently used to assemble the best parts of several different takes of the same song into a composite piece. While basic cut-and-paste editing is possible on a multitrack hard-disk system, a dedicated stereo-editing system or software package will usually include specialist tools that can be used to edit to a much finer degree. Like those in the mainstream multitrack packages, there are also third-party software plug-ins that provide functions such as compression, limiting, EQ, noise reduction, noise-shaped dithering, stereoimage correction and even. three-dimensional sound processing. Most of the current editing packages also include the ability to burn audio CDs with a cheap CD-R recorder. If the CD-R is to be used as a master for duplication purposes, however, the software package must include the PQ encoding (Pause and Cue information) that is used in commercial CDs.

Hard-Drive Practicalities If youre computer,

running a hard-disk recording system on a with or without external hardware, it’s always

better to use a separate drive to store your audio data. Unlike tape systems, hard drives record their information wherever they can find space,

and if you’re

using the same

disk for

word processing, sorting your programs and handling your email, the audio files could end up being spread all over the drive. This process is know as fragmentation, and the more fragmented your drive the slower it gets. Even if you’re using a separate drive for sound editing, the fact that you’re likely to keep adding recordings to it means that it will occasionally need to be defragmented, using specialist software such as Norton Utilities. This software moves all of the pieces of file data around so that the information is stored in contiguous sections of the drive. A separate drive doesn’t necessarily have to take the form of an external box, however; most computers have provision for the attachment of an extra internal drive. SCSI drives are often used for audio purposes because they’re faster than the EIDE drives used inside PCs and Macs, but unless you need a lot of tracks an EIDE drive may well do the job at around a third of the price of its SCSI counterpart. It'll also save you the expense of buying a SCSI card for your computer. If you find that you really need to use a large number of tracks, you should choose a fast SCSI Ultra II or SCSI Ultra II drive and a suitable SCSI interface card. As a rule, the faster

the spindle speed of the drive (measured in rpm) the faster the drive’s data transfer rate. Look for a drive with a high rate of sustained transfer, not just a fast access time, and it’s

always wise to ask the manufacturer of your audio software if they can recommend specific drives. As a rule of thumb, 16 tracks of 16-bit, 44.1kHz audio require a disk with a sustained transfer rate of 1.4Mb per second, and even relatively slow drives can perform better than this. At the other end of the scale, a SCSI Ultra II drive may have a rate

of sustained data transfer in excess of 5SOMb per second. Stereo

material

sampled

at 44.1kHz

(the

same

rate as CD

sampling) uses around 10Mb of disk space per minute, which means that around 600Mb of disk space is required to hold an hour’s worth of stereo material. In practice rather more is required because, if you decide to perform a destructive edit, the system usually creates backup files so that you can change your mind if things don’t work out. In multitracking, the recording time is halved every time the number of tracks is doubled, so if you’re using the same 600Mb drive you’re looking at a 15 minutes of continuous eight-track recording or less than eight minutes of 16-track recording. It’s also worth noting that larger-capacity drives are usually faster, making them more suitable for multitracking purposes. As prices of hard drives are now much cheaper than they used

to be, it’s a good idea to budget for a drive with at least twice as much capacity as you think you'll need.

Recording Hardware The way in which the computer system records your audio material depends on the hardware; budget soundcards are usually equipped with only analogue inputs and outputs, while more sophisticated systems also have S/PDIF and/or AES/EBU digital interfaces. The input and output converters and connectors may be situated on the soundcard itself or mounted within an external box. External box-mounted systems allow more space in which to accommodate connectors,

and

away from computer. Analogue

the

also

allow

electrically

recordings

the

converters

noisy

to be

environment

are fine for multitrack work

mounted

of

the

because

most of the source material is analogue (all you need to do is

run line outs from a mixer or recording pre-amp), but when

editing stereo material it’s generally considered unacceptable to leave the digital domain unless you specifically need to process the audio signal with an analogue device, such as an esoteric equaliser, in which case the digital data which is to be edited is usually fed in from DAT via the digital I/O connectors. A tip here is to buy a proper digital transfer cable, which has a different impedance to that of audio cable; and while you may be able to get away with using a hi-fi phono lead, you might end up with unexplained clicks and glitches.

Choosing A Computer When choosing a computer, you should first decide which software you want to run on it. If you’re not sure, ask other people what they’re using in their own systems. For MIDI/audio recording most people seem to use either Emagic’s Logic Audio or Steinberg’s VST, although in the USA packages such as MOTU’s (Mark Of The Unicorn’s) Digital Performer, Cakewalk and Opcode’s StudioVision are also very popular. All of these packages offer sophisticated MIDI and audio recording and editing capabilities, so you should choose that which best suits your working methods. In some cases the choice of whether to use Mac or PC will be made for you, because the software will only run on one or the other, but some programs are available for both platforms, in which case you can still choose the computer you prefer. Macintoshes are generally easier to set up and use than PCs,

but they cost a littlke more and are not as easy to upgrade. Personally, I think that it’s worth spending a little extra money on a Macintosh, but most people choose PCs because they’re cheaper and more widespread. If you’re buying Mac hardware you know exactly what you’re getting, but PCs come from a variety of sources and not all models are equally

well suited to music work. If you’re contemplating buying a PC-based system I’d recommend that you buy the whole system from the same supplier. If you buy your package as a complete system from a single supplier, it’s up to them to make sure that everything works so that they can’t pass the buck onto another company. You may also get a larger discount if you buy everything from one source.

Hardware Recorders Hardware hard-disk recorders are usually designed to function very much like tapeless tape recorders, although it’s common for manufacturers to include some editing facilities that wouldn’t be available on their system’s tape counterpart. The most important benefit with hardware systems is that they work straight out of the box, and you don’t have to worry

about

whether

your

computer,

software,

soundcard

and hard drive are compatible. They also _ feature conventional metering, multiple inputs and outputs and gapless punching in and out, which works just as it would on a tape machine. Computer-based systems, on the other hand, can suffer from a noticeable amount of monitoring latency, depending on the system. Latency is a very important subject and will be discussed later, but essentially it’s a delay between the point at which a signal being fed into the system and _ its appearance

at

the

system’s

monitor

output.

Latencies

of

several tens of milliseconds are common in cheaper systems, and such a delay can be enough to interfere with timing when overdubbing parts.

Digital Tape Digital tape machines, such as the Alesis A-DAT and Tascam DA series, are simple to operate, with cheap media and a wide user base. These are both eight-track machines, but

because they can be used in multiples they are known MDMs (Modular Digital Multitracks).

as

Two or more MDMs can be locked together when more than eight tracks are needed, although the few seconds it takes for the machines to lock up each time can be frustrating when punching in and out. Little editing is possible, unfortunately, though if a suitable controller is used it’s possible to offset one machine against another so that sections from one machine can copied onto different locations on the other. For example, you may want to copy the backing vocals from one verse onto all of the other verses - which is pretty much the extent of the editing that a tape-based system can perform, and even carrying out this simple task is much more timeconsuming and cumbersome than it is on a computer.

Computer interfaces and software systems are available which allow audio to be transferred from a digital multitrack recorder onto a computer, edited, and then returned to any location on the multitrack recorder to single-sample accuracy. This is clearly useful in situations where tape is the most practical medium for recording (such as when recording large bands or location recording) but where some postproduction work needs to be carried out later.

Tapeless Workstations Hard-disk recording systems require DSP power to process data, and although modern computers are _ incredibly powerful there may still come a time when you need to exceed the computer’s capabilities. The next logical step in this case is to add more DSP power to handle processorintensive tasks such as digital mixing and real-time signal processing.

In

developing

their

ProTools

package,

Digidesign

have

adopted the approach of a computer-based workstation, fitting DSP cards inside the computer or in an expansion chassis. Using their TDM (Time Division Multiplex) buss, data can be routed around the system very quickly, making it possible to create a virtual recording and mixing environment. The system’s only limitation is the amount of DSP power available to it. Other companies also offer hardware solutions, ranging from Yamaha’s cheap SW1000 XG synth/mixer/effects/audio recording card up _ to immensely powerful packages which cost much more than the computer to which they are attached. Unless your requirements are especially demanding, however, a_hostpowered

system based around

a powerful

computer

should

be able to cope with between 16 and 24 audio tracks, provide mixing with basic EQ and still have enough power left over to run a handful of plug-in effects. The latest affordable hardware options are really the modern counterparts of the old cassette multitrackers, in that they provide eight or more recording tracks, along with digital mixing and effects. These systems often offer automated mixing features and the ability to generate MIDI timecode for synchronising to a sequencer, but their most popular feature is that they are a one-box solution to recording and mixing. With nothing more than a few mics and a monitoring system you can start work right away. Some systems can even burn material onto CDs, and because everything is in one box they’re also useful for location recording. The negative side to most hardware systems, however, is that they don’t always integrate with MIDI systems as smoothly as MIDI-plus-audio software, and most can’t be expanded.

Digital Mixers Although a computer is quite capable of mixing signals, it’s sometimes useful to use a hardware mixer on which all of the

controls can be accessed directly on the front panel, and on which levels can be controlled with physical faders. The benefit of using a traditional analogue mixer is its familiar format, with only one knob per function, and the low cost at

which additional inputs and insert points can be added. The disadvantages are that the introduction of automation (beyond that of basic fader and mute automation) isn’t particularly cost effective, and that every time a signal is bounced it has to be converted from digital to analogue and then back again. On a digital system, incorporating a digital mixer means that signals can be kept in the digital domain (thus preserving signal quality), and most digital mixers provide the ability to automate EQ, pan and level, as well as aux send levels and

even types of effect. On simpler systems the automation may be in snapshot form, where each snapshot comprises a series of console settings that can be switched in at the appropriate point in a song, although on more upmarket models it’s usual to have fully dynamic automation controlled via an internal automation system, on either an external computer or a MIDI sequencer. The majority of digital consoles also include effects such as reverb and delay as well as dynamic processing, such as gates and compression. This mitigates the lack of insert points to some extent, although it does restrict your

options

when

you

want

to

use

external

effects

and

processors. The downside of using digital mixers is that it’s seldom practical to provide one knob per function, and so the operating system is invariably less intuitive than that for an analogue outputs

mixer. Furthermore, and insert points

additional analogue require expensive

inputs, A-to-D

(Analogue-to-Digital) and D-to-A converters, and so cheaper models are likely to have quite limited analogue I/O

facilities. Connection to computers or hardware digital recorders is likely to be via an A-DAT lightpipe or a similar multichannel digital interface, although newer and faster systems based on FireWire technology are likely emerge. In a small studio system where synchronisation is achieved via word clock, the mixer is likely to be used as the word clock source.

Stereo Recorders Even though a lot of today’s music is generated inside the computer, there nevertheless remains a need for hardware stereo recorders, such as DAT machines and MiniDisc

recorders, both of which are available with digital and analogue ins and outs. DAT machines are widely used in both project and professional studios, largely because of their widespread usage and the convenient media. DAT tapes are like tiny video cassettes that can hold up to two

hours

of

audio

at

16-bit,

44.1/48kHz

resolution,

although most professionals are wary of anything over 90 minutes in length. Recording is straightforward, as long as you ensure that the input level never hits the clip LED at the top of the meter. Because DAT uses a rotating-head system like that of a video recorder the actual tape speed is extremely slow, so it’s usual to leave between 30 seconds and a minute of tape blank at the start of the tape just in case any dropouts occur near the start. Unlike analogue tape, however,

DAT machines

sample clock, should record wind to the Similarly, for

record a subcode derived from the

even when the audio input is silence, so you silence from the start of the tape rather than one minute mark and record from there. the time counter to work properly the subcode

must be continuous,

which

means

that, when

you want

to

add more material to a tape onto which you’ve already recorded material, you should wind back to a point at which

the previous subcode is recognised before you resume recording. Clearly this means allowing the recorder to run on for a little after each recording, in order to lay down a few extra seconds of subcode. Failure to maintain a contiguous subcode in this way will confuse the real-time counter - for example, it may start again from zero after a section of unrecorded tape.

MiniDisc MiniDisc is a consumer audio format that uses an aggressive data compression system to reduce the amount of data needed to represent a given length of audio. Essentially, anything that the compression algorithm thinks is rendered inaudible through masking by other parts of the music is discarded. Because of this compression MiniDisc is not as accurate as DAT, but it’s still good enough to use for demos or independent releases of pop music with a limited dynamic range. Indeed, most people can’t distinguish it from CDs, although a discerning listener can usually spot the difference, especially on classical material. I would recommend mixing onto DAT if possible, but MiniDisc will suffice for all but the most demanding applications.

SCMS SCMS (Serial Copy Management System) was introduced after the record companies pressured manufacturers into producing a device which could protect their works from piracy in the digital domain. SCMS is implemented by almost all consumer digital recording hardware, and works by inserting an invisible flag in the data stream. Depending on the status of this flag, the recording you make can be copied freely, copied for one generation (so that the copies can’t be used to make further copies) or copying can be prohibited altogether. A source recording can be copied digitally but a

flag is inserted in the new copy which will prevent any machine fitted with SCMS from making any further copies from it. Of course, you can make as many copies as you like from the original, and you can still copy in the analogue domain, so the system is actually more of a nuisance to bona fide studio users than it is a deterrent to pirates, as you usually end up with ‘copy prohibited’ flags on copies of your own material. Fortunately, most computer-based editing systems strip out the SCMS flag.

CD-R Now that blank CD-Rs are cheaper than analogue cassettes, they have become very attractive for those wishing to archive audio data from sequencers as well as for storing final stereo mixes. Computer-based CD-R recorders are fairly easy to use if they are issued with decent software, and the discs are reasonably durable as long as they are handled carefully and stored

in a dark,

dust-free

environment.

You

should

avoid

cutting corners by using cheap, unbranded discs, as you may end up losing valuable data, and try not to get dust and fingermarks on the discs before recording onto them, as this can cause problems during the burning process.

Now that it’s possible to record, mix and master material without it ever leaving your computer, it’s possible to burn a CD directly without having to master your material to DAT or MiniDisc. Rewritable CD-Rs are also useful for recording audio data from your sequencer, but they’ll normally only play back on a CD writer, not a computer CD-ROM drive or audio CD player. However, some of the new generation of consumer CD players will work with CD-R, and it’s possible that future models will do this as standard.

3 THE SOUNDCARD

STUDIO

In the past music studios were complex and costly affairs, comprising open-reel tape machines, synchronisers, huge mixers,

MIDI

sequencers,

sound

modules,

keyboards,

outboard effects and miles and miles of cabling, but the possibilities offered by the computer desktop studio means that it no longer has to be this way. Computer soundcards have received a lot of bad press in the past, mainly because of the very cheap models that are supplied as standard with PCs, enabling their owners to play games. Fortunately, a number of serious manufacturers have turned their attentions to designing soundcards, and have produced models on which the audio converters do justice to 16- or 24-bit audio and, where provided, on-board synthesis capability is derived from either existing keyboard instruments or sound modules which cost three or four times as much. What’s more, because PC soundcards are purely software-driven devices, there is often a surprising amount of useful support software bundled into the card, including sequencers, editor librarians, MIDI song files and mixer maps for the most commonly used sequencers. As mentioned previously, choosing a card with multiple inputs and outputs is likely to cost more than a stereo in/stereo out card, but even multichannel cards start at a very reasonable price when you consider how much open-reel tape recorders used to cost. As a very general rule, the soundcards that are also capable of synthesising and interfacing with MIDI have evolved from

the games market, and usually offer stereo analogue ins and outs, with the better models also offering S/PDIF ins and outs. Multiple input and output systems are more often designed to provide audio-only support, so you may need to look elsewhere for synthesis and MIDI interfacing. Because PCs have a number of expansion slots it’s possible to use two or more soundcards in the same machine, although installing hardware isn’t always the painless procedure it ought to be, particularly if the cards are of different types. In most cases it’s probably best to choose a single card that provides the necessary features, or to use two cards of the same type to increase the amount of available I/O. The newer PCI cards used with Windows ‘95 are relatively easy to install, but you can still run into trouble if you’ve never delved into the nuts and bolts of computer operating systems before, so I make no excuses for once again advising you to buy a fully installed and configured system if at all possible.

Macintosh Users Both Macs and PCs now use PCI card slots, so both systems can use soundcards in much the same way, although, unlike PCs,

all

modern

Macs

come

with

basic

16-bit

audio

interfacing built in. Stereo inputs and outputs are accessed via mini jacks, which can be selected as the audio input and output devices via the Sound control panel of the Mac. If you have any other Sound-Manager-compatible audio hardware installed, this control panel will allow you to select between Mac internal hardware and the installed hardware. Although the sound quality produced by these internal converters isn’t as good as that produced by a high-quality audio interface or a dedicated card, it’s still good enough for serious demo work.

Connections While a cheap soundcard plugged into a pair of nasty little powered AV monitors isn’t going to threaten’ the conventional studio, the more serious user can gain access to

some pretty serious music power on a computer for very little additional expense. What’s more, it’s often possible to mix and process sounds in a far more comprehensive manner than you might imagine. For example, there may be several separate stereo outputs running from different soundcards and daughterboards that can be conventionally mixed with an external analogue or digital mixer. While there’s usually an option to daisy-chain the audio ins and outs of various soundcards to enable you to mix everything into stereo, the quality of the sound drops off significantly if mixed in this way, and by keeping outputs separate it’s possible to patch into an external mixer, which is far more flexible.

A digital studio doesn’t have to be completely digital. In a simple setup a small analogue mixer is the most cost-effective model to use, because it provides high-quality sound, you can patch in outboard effects if necessary, and you don’t have to worry about sample rates and synchronising if you’re looking to add material from other digital sources, such as CD players, DAT machines and MiniDisc recorders. Keeping a few mixer inputs spare is always a good idea, as sooner or later you'll be tempted to add a hardware synth or a sampler to your system. Monitoring can be carried out on a proper monitoring system

or a hi-fi rather than on cheap, powered desktop speakers, and with the improved audio quality provided by an external mixer you can produce a very respectable mix, especially if you master to DAT or MiniDisc rather than analogue cassette. There are also other reasons why adding a mixer is often a good idea.

While PC soundcards often come with mic or line inputs that allow you to record your own .WAV audio files (the standard PC sound file), the signal path through these is usually pretty poor and they have no provision for connecting professionalquality studio mics. A much better option is to use an external mixer or mic preamp to bring the output from a decent mic up to line level and then feed this into the soundcard’s line input. Not only does this allow you to use a studio-quality mic but it will also improve the signal-to-noise ratio of your recording by amplifying the signal within the mixer, rather than within the soundcard. Even if you’re using a fairly basic general-purpose mixer you can use one channel as a mic pre-amp simply by turning the channel fader right down and using the channel insert send to feed your soundcard. The mic gain is set with the channel’s gain trim facility, and the signal level is monitored with whatever facilities are provided with the soundcard. Figure 3.1 illustrates how a typical PC-based system might be connected.

Monitor speakers

Pawer amplifier Mixer monitor

out fo power amplifier (main

Seema] | foe's sae external hardware

feseartherh

aymihs and other 7 * SOurceS

Recording pre-amp feeds mic signal inta soundeard, | mixer could also be

used as 4 pre-amp by wing the channal inser sand of a

Prt Tn ae oe ON

pre-fade aux sand to feed the soundeard 4

Small mixer

Soundcard

audio line input

Joystick pert (MIDI) MIDI adapior

lead plugged into joystick

MIDI Out MIDI

PC computer

Audio Gut

from synth

Keyooard and manitar

Figure 3.1: A typical soundcard studio

If your mixer has no insert points, you can use the pre-fade aux send on the mic channel to send the mic signal to the correspondingly numbered aux-send output and use this to feed the soundcard input. For example, you could feed your mic into channel one and route this to the aux one output by turning up the aux one send on only channel one. The level fader on channel one would be off, and the level would then

be sent to the soundcard controlled by the aux one master output level knob. Outboard effects may be connected to the mixer in the conventional way via post-fade aux sends and effects return channels. If you’re not familiar with this aspect of mixers, the book basic Mixers - also in this series-investigates them in some depth. Alternatively, you could consult your mixer’s manual.

Recording Pre-amps A number of companies produce recording pre-amps that combine a high-quality mic/line/instrument pre-amp with a compressor and an equaliser. These plug directly into the line input of a soundcard and enable you to record with a highquality studio mic. Electric guitars and basses may be plugged into the high-impedance instrument input, while line-level sources, such as keyboards, may be recorded via the line input. The compressor produces a more even sound, and compression is invariably used on vocals in commercial recordings. With a high-quality equaliser you can adjust the tone of the sound before it’s recorded, and if you prefer to record one track at a time then the best way of ensuring the optimum sound quality is to use a recording pre-amp, as the mic section on a pre-amp is generally better than those found on low-to mid-priced mixers.

And There’s MIDI A basic PC MIDI system will generally use the MIDI interface provided by the soundcard, and the standard way of working in this case is to use a MIDI adaptor cable, which plugs into the card’s joystick port. Simple one-in/one-out MIDI interfaces are limited in that you can drive only 16 external MIDI channels from them. With

today’s synth modules, this usually means a _ single multitimbral instrument, but you should remember that all internal soundcards (except the daughterboard) use virtual MIDI ports to avoid tying up the external MIDI port. This is obviously good news if you’re on a strict budget, as you can then use the sounds present on your internal soundcard at the same time as one multitimbral external MIDI module and still only need one MIDI port. If you need more ports in order to handle additional external synths then you can buy multiport MIDI interfaces, as either cards or external hardware boxes, which offer eight or more sets of MIDI outputs. These interfaces should not be confused with simple multi-output Thru boxes, however, from which each MIDI Out carries the same information - a multiport interface’s outputs are quite independent, and are usually identified by numbers or by letters. For example, port A would carry MIDI channels 1-16A while port B would carry channels 1-16B, and so on. An eight-output multiport interface provides 128 MIDI channels, all of which can be addressed separately, and this should be sufficient for all but the most equipment-intensive arrangements. It’s important to

check that the interface you are considering is designed to work with the audio software which you intend to use. A ‘dumb’ MIDI master keyboard can be simply connected to the computer’s MIDI In, while a MIDI synth keyboard would need to be set to Local Off so that the synth section could be driven from the MIDI Out port of the computer. Figure 3.1 also shows the configuration of this latter option.

Evolution The PC-or Mac-based MIDI/audio studio has the advantages of being relatively inexpensive, compact and (once set up) convenient. With the addition of a mixer, a monitoring

system and possibly some outboard signal processing you should have the basis of a serious desktop music recording system at your fingertips, and as soundcards continue to become more powerful and better specified it’s possible to upgrade a system a piece at a time without having to sell up and start from scratch. However, it’s wise to buy the fastest computer you can afford because, as new software versions become available, they always include new features that make more demands on your processor. By choosing your soundcard or audio interface carefully, and incorporating a small mixer along with a modest amount of external signal processing, it’s possible to build a serious desktop studio for less money than stereo samplers used to cost.

Software Synthesis Software-based

synthesis

(ie that which

uses the computer’s

own processing power to create sounds, rather than hardware) is becoming more practical now that highpowered computers are relatively cheap. Obviously, these packages use up some of the computer’s processing power and memory, and this can be anything from just a few percent up to nearly the entire processing power of the system, depending on the software’s sophistication and on how well written it is.

Latency And Monitoring Latency is one aspect of working with the soundcard studio around

which

manufacturers

tend to skirt, and often users

don’t understand its full implications until they’ve already parted with their money. Latency is that infuriating delay which occurs between the sound being fed into a computerbased recording system and re-emerging from the audio outs. Latency arises when audio is routed via the host processor (which is necessary if you wish to monitor the sound while

software plug-in effects are running), and because of the way in which the computer operating system handles data this round trip can take a noticeable amount of time. You can also hear the effects of latency when you hit the Start button to play back a sequence, as there is often a short delay before playback actually begins. On a well-designed system this latency may only be a few milliseconds, in which case few people will even notice it’s there; but other combinations of hardware and software can

throw up delays latency is only a whatever you're added. This can great. There are minimising

of more than half a second. As a rule, serious problem when overdubbing, when trying to overdub comes back with delay throw out your timing if the delay is too

both hardware and software approaches to latency, notably Steinberg’s ASIO II drivers

(software) and soundcards with integral DSP effects and thru

monitoring (hardware). Thru monitoring simply means that, when overdubbing, the audio passes straight through the soundcard and onto the monitor outputs without going through the computer’s CPU. But if you already have a system and latency is giving you trouble, what can be done?

Input Monitoring The answer is to use a mixer to set up the monitoring. Unless you have an extremely simple system, and you rely on the same soundcard to provide synth sounds and audio recording, you'll probably already use a mixer to combine the output from your soundcard with the outputs from your external synths, samplers and drum machines. Furthermore, unless you’re going to trust all of your audio processing to VST plug-ins or similar computer-powered virtual devices, you'll need a recording pre-amp with an integral compressor,

which will even out vocal levels during the recording stage. Assuming

that you’re

in possession

of this equipment,

you

can now set up a simple system in which the output from your recording pre-amp is split to feed both the mixer and the soundcard input. At it’s simplest this could be done with a Y-lead (it’s always all right to split signals this way but not to mix them), but you'll find that many recording pre-amps already have two outputs, often one on a jack and the other on an XLR. If this is the case, you can feed the jack output to the soundcard and then make up a_balanced-XLR-tobalanced-quarter-inch jack lead to feed the mixer’s line input. You might be able to get away with using an XLR lead plugged directly into the mixer mic input, but only if your mixer has a pad switch and plenty of attenuation range on the input gain trim. If you don’t have a recording pre-amp you may instead be able to use one of your mixing channels, as long as it has a direct output (or insert send) to feed to the soundcard input.

This will allow you to monitor your overdub directly through the mixer when recording, but you’ll need to mute the output from the track being recorded within your software or you'll hear both the direct monitor signal and the delayed signal. Some packages have a thru monitoring on/off switch, while with others you simply have to use the software’s virtual mixing facilities to turn down or mute the output from the track being recorded. After you’ve recorded your material you’ll need to unmute the track output before you can hear it (if you killed the delayed monitor signal by muting it, that is), and you may also want to mute the mixer channel on which you monitored your recording pre-amp or your live mic will still be feeding the mix. However, this system is limited in that,

because

you’re

bypassing

the

computer

when

monitoring

overdubs, you won’t be able to hear the effects of any VST plug-ins until after the track has been recorded. If you need to apply reverb to help the singer provide accurate vocals, you can still hook up a regular hardware effects unit to your mixer and add as much monitor reverb or echo as you like it won’t be recorded. The complete setup is shown in Figure 3.2.

Punching In If you’re executing a punch in to patch up a mistake in your recording, using thru monitoring means that you won’t be able to hear the original track, as it will be turned down or muted. A practical solution to this problem is to record the punch-in part on a new track and then paste it into the original track once you’ve got it right. It’s easier to edit vocals if you can punch in and out between words.

Choosing A Soundcard The first things you should decide is whether or not you need a stereo or multichannel card. How many tracks do you need to record at one time? Do you need to get them out of the computer separately, or do you plan on mixing internally? Applications like Cubase SX and Logic Audio are powerful in that they allow a number of audio tracks to be recorded separately and then mixed down. The mixing is carried out inside the computer, and so is possible with just a stereo output.

Mixer Audio Out —p— to maniors

Recording pre-amp

_ fic input

i, Solit lead or

| | second audia Oubpuy

Soundcard output Soundcard When recording. ihe @UIDUE

Teor he

input

pre-amp

is Monitored directly through the mixer

Master keyboard

Figure 3.2: Thru monitoring via a mixer

Many soundcards now offer better than 16-bit resolution, but both 20- and 24-bit recordings take 50% more hard disk space than 16 bit. 18-, 20- or 24-bit converters are used because they provide more dynamic range than 16-bit converters, but unless the rest of your studio equipment is of superb quality these extra bits will probably record only the hiss from your mic pre-amps or traffic noise. If the card you’re looking to buy offers these higher bit depths then by all means try them to check if you can hear a difference, but if you’re recording regular pop music you'll save a lot of hard drive space by switching to 16-bit operation. If you need to send audio from your computer to a digital

stereo recorder, such as a DAT or MiniDisc machine, you'll need a card with a digital output, and it’s best to get a card with both digital ins and outs to provide you with more flexibility. The most common consumer digital interface is S/PDIF (Sony/Philips Digital InterFace), and there are two types available: the more common is co-axial (which uses phono sockets), but there is also a version, known as Toslink,

which uses fibre-optic cables. I mention this because you need to ensure that the S/PDIF connector on your soundcard matches that on your recorder. There are adaptors available

which convert expense.

between

the two,

but this is just needless

Professional audio equipment uses AES/EBU, a balanced system which uses XLR connectors, although these aren’t used on budget soundcards. S/PDIF will prove to be adequate in most instances.

After deciding exactly how many input and output channels you're going to need, you should look at the other features you may need, such as on-board synths, MIDI interfaces, hardware mixing and effects. Furthermore, you should also find out whether daughterboards can be added later, thus increasing the synthesis capabilities of the card. If you choose a card that can handle sampling, make sure that there is provision for adding more sample RAM (unless the card is designed to use the computer’s own system RAM).

Soundcard Drivers You'll need a piece of software called a driver to make a soundcard work with your audio software. Cards are useless without the right drivers, and the quality of the driver can dramatically affect the amount of latency you have to endure. The driver is usually part of the package that comes with the card, but if the card happens to be made by the

same company that makes your audio software you might find that the audio software has the necessary driver already built in. It’s also quite common for driver software to be updated in order to improve its performance or to iron out any bugs, and the best way to obtain an up-to-date driver is to make frequent visits to the card vendor’s web site, where the most current downloads are often available. The most popular audio software is compatible with Steinberg’s ASIO (Audio-Specific Input Output) driver standard, so if you want to use this you should make sure that ASIO drivers are available for your card. And if you find all of this stuff about drivers a little heavy, then that’s just one more reason to buy a complete, working system.

BeOS Most PCs run on Microsoft’s Windows software, but this isn’t

ideal for music programs because it’s designed as a broad platform to support a number of different types of software, from word processing and Internet browsers to graphics and spreadsheets. Be Corporation have developed an alternative operating system that runs on suitably specified PCs and on pre-G3 PowerPC Macintoshes. This operating system is known as BeOS (Be Operating System), and is specifically designed to handle multimedia work effectively and with very low audio latency. The catch is that software must be designed to run on BeOS - you can’t just load up a Windows program. Fortunately, several major audio software manufacturers are already developing versions of their programs for use with the BeOS. At a recent demonstration I tried a software synthesis package on both BeOS and Windows

‘98, and while

the Windows

version

had

around

half a second of latency, rendering it unplayable, the BeOS synth responded so quickly that it was just like playing a hardware instrument. A low latency is obviously very important with software instruments, because any

appreciable delay will prevent you from playing in time. The range of music software that will run on BeOS is for the moment fairly limited, but from what I’ve seen so far the BeOS system could make the PC a much better platform for audio than it is currently with Microsoft’s Windows software.

No Games! The secret of having a smooth-running PC-based studio is to use it for only music. If you picked a PC for the kids to play games,

or to surf the

Internet,

you

could

be

heading

for

trouble. Games often take over certain resources and operate in unorthodox ways to get the maximum performance out of the game, and you may find that they mess up the way in which your audio software runs. If you need to use a PC for two distinct areas of work, it may be best to get a dual boot drive system with a separate copy of Windows installed on each drive and switch between them, using one drive exclusively for music and the other for all of your non-music applications.

On the Mac OS it’s possible to select your start-up drive. However, even a dual boot system may not be necessary, as you can use the Extensions Manager control panel to create a customised set of extensions specifically designed for music

use, with all unnecessary extensions switched off. In this way you don’t need two versions of the Mac OS on different drives - you can simply start up with the appropriate set of extensions. Again, if you’re new to this aspect of computing, ask a computer specialist (or a friend who has already gone through this) to help you out.

4 MUSIC SOFTWARE

No matter which computer platform you finally choose, the ultimate functionality of your system will be determined by the software you buy to run on it. Music software tends to fall into a number of main categories: MIDI sequencers, MIDI-plus-audio sequencers, multitrack recorders (which are much like sequencers but with no MIDI capabilities), stereo editors, effects and processor plug-ins, and virtual synths. There are also programs that fall between the serious sequencer and hobbyist categories, most of which function by using and processing existing musical loops and sound samples.

The Sequencer A modern MIDI sequencer will provide a bare minimum 16

MIDI

some

tracks

level

of

(often

audio

many

more),

support.

and

With

most

a

now

MIDI

of

include

sequencer,

numerous separate musical parts can be recorded at different

times, either through the parts being played one at a time on a MIDI keyboard, through note and timing data being entered manually, or through a combination of ‘live’ playing and editing. Once recorded, these parts may be played back via any MIDI-compatible synthesiser or collection of synths. You'll also need a synthesiser (either a keyboard instrument or a MIDI

module),

unless

you

have

a soundcard

with

one

already built in, and the number of different musical parts that can be played back simultaneously depends on the number and type of the synthesisers that are at your disposal. Fortunately, most modern synthesisers and PC soundcards

are capable of playing back a minimum of 16 different sounds at once, each controlled by a different MIDI channel. This ability is known as multitimbrality.

The MIDI Composer Although the MIDI musician uses a lot of electronics, the process of composing with MIDI isn’t that different from the traditional approach to composing. Like orchestral composition the work usually starts at the keyboard, but in this case the keyboard is a MIDI instrument connected to a MIDI sequencer, rather than a piano. Instead of writing down a score the composer records sections of the music into the

sequencer against an electronic metronome set to the desired tempo, and instead of scanning a score to verify what’s been done the composer simply plays back the recording on a suitable synthesiser in order to hear exactly what’s been recorded. A multitimbral synthesiser can play several musical parts at once, and when used in conjunction with a sequencer

each ‘part’ of the multitimbral synth can be used to play back one musical part. For example, you might have two sequencer

tracks dedicated to strings, one to horns,

one to

percussion and one to woodwind.

MIDI And Sequencing It’s often convenient to visualise a sequencer as being analogous to a multitrack tape recorder, but it’s vitally important to remember that, with MIDI, what is being recorded is not the sound itself but the electronic equivalent of a musical score. Just as a musical score is a series of instructions

to the musicians,

a MIDI

sequence

contains

a

series of instructions which tell your synths what to play and when to play it.

Audio Sequencers

The audio tracks within a MIDI-plus-audio sequencer record digitised audio directly onto your computer’s hard drive. These tracks may be edited in much the same way as MIDI tracks, so the ability to manipulate MIDI and audio together within the same environment is very attractive if you enjoy

experimenting

with

different

arrangements.

Unlike

tape

recording, MIDI tracks can be divided into sections, and these

sections can then be moved to different locations in a song. Furthermore, the same sections may be used two or more times without re-recording the part, so if you want to you can use the same backing vocals on every chorus. Similarly, a short section of drums can be looped and repeated to provide

a continuous drum track.

Sequencer Setup In a typical setup, a master MIDI controller (usually a keyboard, but not always) is connected to a sequencer with a MIDI cable, and when the sequencer is set to record, any notes played by the keyboard are recorded as MIDI data into whichever sequencer track has been selected for recording. A simple system might have 16 MIDI tracks set up so that each is on a different MIDI channel, and if the MIDI output of the sequencer is fed to a 16-part multitimbral module all 16 tracks can then be played back at once. If you only have an eight-part multitimbral module, however, only eight different sounds can be played back at once, just as a real-life eightpiece ensemble can only play a maximum of eight different lines of music at the same time. (Figure 3.1 in Chapter 3 shows

how

a

typical

computer-based

sequencing

system

connects to an external keyboard synth.) If you have a keyboard that includes a synth, as shown here, simply select Local Off and connect it up like any other synth module. Local Off isolates the synth’s keyboard from its soundgenerating circuitry so that, in effect, it behaves as if it were a separate dumb keyboard and MIDI synth module. This is

necessary in preventing MIDI information from being fed around the system in a continuous loop, which usually causes trouble with stuck or repeating notes.

Metronomes Although a sequencer can be treated as nothing more than a multitrack MIDI recorder, its real power lies in the way in which it can modify or edit recorded data. When a recording is made, the sequencer is set to the tempo of the desired recording and a metronome click is laid down so that the musical performance can be synchronised with the internal tempo of the sequencer. In this way the MIDI data is arranged in musically meaningful bars, which makes editing note timing or copying and moving whole sections a very easy and precise procedure. With music incorporating tempo changes it’s possible to enter

a new tempo at any bar or beat location, although some budget sequencers may impose a limit on the number of tempo changes which can be placed in each song. More sophisticated sequencers may even have a graphic Tempo Editing mode, in which smooth increases or decreases in tempo can be created by drawing slopes. If you don’t want to be tied to a tempo at all, you can simply turn off the metronome click and play normally.

Tracks And Channels At this point in the proceedings it’s very easy to mix up MIDI channels and sequencer tracks, but they’re very different. A sequencer track is simply somewhere to record one layer of your composition, but the MIDI information in that track can be on any MIDI channel you want. A track can even contain

MIDI data relating to two or more MIDI channels, although mostly a single track records data on a single channel to

avoid any confusion. It’s also possible to have several different tracks recording MIDI data set to the same channel. For example, if you’re recording a complicated drum part you might want to lay the bass and snare drum onto one track, the cymbals and hi-hats onto another and any tom fills onto a third. All of the drum sounds may be on the same MIDI channel, but because they’re on different tracks they’re effectively recorded as different layers. Not only does this make the parts easier to play but it also makes it less confusing when editing in any changes. Figure 4.1 shows the Arrange page of a popular computer-based sequencing package, depicting the layout of the tracks and the way in which recorded sequences are represented. ffa=—

S Mem ape lien

SG Shere § Ser a PR

Pees!

2

© ters

6° Pees

cum

7

Pee

a Spee)

:

yt

ar

3 Bere

= were

1

= WB

rei

fa

Citi frac ;

5 GRE rrr

=

Peles) CM 000

Figure 4.1: Sequencer Arrange page

Cut, Copy And Paste The remaining capabilities of a MIDI sequencer bear more resemblance to a word processor than anything else. Like a word processor, you can delete or replace wrong characters (in this case musical notes), and if you want to use the same

phrase more than once you can copy it and paste copies of it into new locations rather than having to play the same material several times. MIDI information comprises not just note information but also controller data from modulation wheels, pitch-bend wheels, expression pedals and so forth. Unless you deliberately filter out certain types of MIDI data (and some sequencers have the facility to do this), you’ll find that your sequencer

captures

note

on,

note

off,

velocity,

pitch,

modulation and aftertouch data, along with other controller information and MIDI Program Changes and Bank Change messages. When recording a part that needs the addition of a lot of complicated pitch bending or vibrato, a useful trick is to record the part straight onto one track and then record the vibrato and pitch-bend wheel data on another track set to the same MIDI channel. As you record the controller data track you'll hear it affecting the performance on the original track. A sequencer track must also be informed of the synthesiser sound which it is expected to control, so in addition to the MIDI channel number (which tells the sequencer which instrument or which part of a multitimbral instrument it is controlling), it’s also necessary to enter the program number of the patch you want to hear, and if the synthesiser supports

MIDI Bank Change messages you'll also need to tell it in which bank the sound is located. For this reason it helps to photocopy the relevant patch lists from your synthesiser manuals and pin them to the wall next to your sequencer. If you're not very experienced with MIDI, the basic MIDI book in this series covers this area in greater detail.

Playback When a MIDI sequence is played back over a synthesiser, the sequencer transmits the MIDI information to the receiving

synth in exactly the same order and with exactly the same timing as it was originally played, although it’s possible to change the tempo after recording without affecting the pitch (unlike a tape recorder, which uses analogue sound rather than MIDI data). The level of each part in your composition can also be controlled with MIDI, usually by changing the value of MIDI controller seven, which governs master volume. Most sequencers provide an easy way of doing this.

Editing On the editing pages of a typical sequencer it’s possible change the value, start time, length and velocity of any of notes you’ve played, or you can build up compositions entering the notes manually, placing new notes onto quantise grid in non-real time, in much the same manner writing out a manuscript.

to the by the as

If you’re working with a package

that has a scoring facility, it’s also possible to enter notes directly onto the score, almost as though you were writing notes on manuscript paper. The entry of note information in non-real time is sometimes known as step-time entry.

Quantising An important feature common to both hardware and software sequencers is the ability to quantise data after it has been recorded. This is useful for those users who don’t have a perfect sense of timing. Essentially, when you choose to quantise something, the timing of your piece is changed so that every note you’ve recorded is pushed to the nearest exact subdivision of a bar. For example, if you’re working in 4/4 time and you select 16 as your quantise value, every note moves to the nearest point on an invisible grid dividing the bar into 16 equal time slots. The quantise function must be used carefully, however, as it can strip all of the ‘feel’ from some types of music; however,

if youre working with dance music, with timing is essential, the quantise function is Remember that the quantise function will meaningful results if your original recording time with the sequencer’s metronome click.

which precise indispensable. only produce was made in

MIDI Drums Most soundcards and synths have built-in drum sounds, and traditionally these are often assigned to MIDI channel ten. Using drums is exactly the same as using any other keyboard sound, except that, instead of having the same sound playing at different pitches across the keyboard, each key will play back a different drum sound.

Types Of Sequencer All MIDI sequencers use computer technology, but you can choose between buying a sequencer system that runs on an

existing computer (such as an Atari ST, Apple Macintosh, IBM PC or Commodore Amiga) or opting for a piece of dedicated hardware, on which everything you need is built into one box. These two groups work in a similar manner, and tend to vary only in the way in which the recorded information is displayed and how easily it can be edited. Hardware sequencers are also built into workstation-type keyboard synthesisers, and some hardware sequencers also have built-in synth modules. Hardware sequencers offer relatively accomplished players the benefits of simplicity and convenience, but because they can’t display as much information as a full-size computer screen, and because there’s no mouse, editing is generally less comprehensive and more time consuming than it is on a computer-based system. However, recording usually involves little more than selecting a track, hitting Record and playing the material. Another benefit of hardware

computers is that

they are more practical in live performance situations - they are more compact and more rugged than a computer and monitor, and there’s also not as much to plug in.

MIDI Data Storage It’s one thing to record a MIDI sequence, but what do you do with it once

it’s finished?

There’s

no manuscript

paper

on

which to store your work; song data is instead stored as a MIDI song file on a floppy disk or hard drive. Computerbased sequencers lose their stored information when they are switched off, so it’s vital to save your work to disk at regular intervals. It’s also an unfortunate fact that computers crash just when

you least expect them to, so it pays to save your

work every few minutes.

Computer Complexity Computer-based sequencers are capable of considerably more sophistication than most hardware models, which means that they often have a steeper learning curve. You must always familiarise yourself with the general operation of the computer before trying to tackle a sequencer package, although in my opinion this is more than compensated by the amount of visual control available, especially when it comes to creating new song arrangements or editing recorded material.

The MIDI Interface On a hardware sequencer you simply plug your master keyboard into the MIDI In socket, plug a synthesiser into the MIDI Out socket and you’re ready to go. Computers, on the other hand, don’t usually have MIDI sockets - the obvious exception being the popular but now-ageing Atari ST. This means that, unless you’re using one of these, you'll need to buy an external MIDI interface, one of the synth modules that

comes with a built-in MIDI interface, or a soundcard that has MIDI In and Out ports.

MIDI interfaces for Macintoshes plug into the modem or printer ports on the back of the computer, while PC users need either a soundcard fitted into their computers or an external plug-in interface. Most PC multimedia cards include a MIDI interface facility, although it may be necessary to buy a special adaptor cable to make use of it.

The User Interface Most of the market-leading software sequencing packages have adopted the style of interface which was pioneered by Steinberg in their Cubase software. The success of this interface is based around the fact that it uses a digital analogy of multitrack tape. The sequencer tracks are depicted on the computer screen as individual vertically stacked strips, with musical bars running sequentially from left to right. Once a section of a track has been recorded it is shown as a brick running from the start location to the end location. This sequence may then be dragged with the mouse to a new position in the same track, or it may even be moved to a completely different track so that it plays back with a different sound. Sequence blocks may also be copied, split into shorter sections or deleted as required. Most software sequencers comprise a main page, on which to

handle basic recording and arranging, along with a number of other pages addressing various aspects of editing and, where applicable, scoring. The Record and Playback controls are invariably designed to look something like a tape recorder’s transport control buttons, and the Edit pages usually allow you to examine and change the recorded data as a list of MIDI events either graphically, as a display resembling the roll of a player piano, or as a conventional

music score. Most sequencers also have graphic editing capabilities for editing controller information. Figure 4.2 shows some of the Edit pages from a popular software sequencer. Some software sequencing packages also _ include sophisticated score-writing facilities, which enable you to print out sheet music of your compositions, although to achieve this you’ll need a printer which is compatible both with your computer and the software package. Also, because the computer doesn’t always interpret what is played in the same way that a trained score writer would, some musical literacy is also useful.

fa fe ]emt Functions Wiew| se

‘ai ee

PEE

ens

la

Meu Age ld..Etherialt copied Pa

i

aAbica

| ------------- Stert of List -------------

wee

2ieod1lii “i oii i 2ii1i1iit. fied o& fi ol i 2s od o25 1 it i en a | 25 1 it 25) 61COd 20 41 st zo 1 it 20 #41 #& ah od 60 53 list

Ln fm|tatt Pamcticns tirell —

a

NOTE AOTE NOTE NOTE WOTE NOTE WSOTE ANOTE ASOTE AWOTE NOTE MOTE AOTE MOTE ANOTE kw

ruc

"

Ll iL i i L i Li L L iL Ll i iL i L

Bl E? GH? CHS O83 GFL E? G8? DHS 683 FHI off CHS ofi CHF

age ho... iat

45



1 1 1 1 i 1 L 1 1 1 1 1 1 1 1

as

‘.

=

i

3 3. 3 3

3 3 3 & 35 3% % 5 3 3% 30Ud4UCUCUG 53 FF 5 35 53 3 3 5 53 3 3 5 53 3

228 2if O88 Lee 68 252 2352 CS F 2352 3 232 3 S52 4& Se §& S52 5 Se 3 S52

7

rs

=

=



maak

,

1 1d 1 4 1 & 3

HO econeed ; nee

Le

—_

25 42 12 iL? 16 103 183 iff 183 i183 OF 4? OF WY? OF

.

E |=]

az

a

i

=

+.

me]

ok

sl

(=) ee —r_|

|

y=,

aad

|

a

5

=—

i

o

]

i) (

——

ral: :: fae

= =) |

| co

| |

a

——-

a a

5

i

elie

| i

iG

p35

—, =e a pe Lace

Hew Age id, cb ‘ora Feasee

Alfie

Tet Shine

SOE

ations teas

i |

record

EEA

7 i

42.

$3 a

eee i

tr al

SS

43

B eae

ie

44

ee

= oF

2 Figure 4.2: Sequencer edit pages

Extra MIDI Ports Basic MIDI interfaces provide single MIDI output sockets, which means that a maximum of 16 MIDI channels are available to the user. However, you may wish to use two or more multitimbral synthesisers to create a composition with more than 16 parts, or - as is more often the case - you may wish to change between a number of synthesiser modules

without

having

achieved

by

to

using

reconnect a MIDI

MIDI

interface

leads. with

This

can

multiple

be

output

ports. It’s impossible to have than 16 MIDI channels, but if a multi-

port MIDI interface is used in conjunction with compatible sequencing software it’s possible to have several different sets of 16 MIDI channels. The ports may be designated by number or letter within the sequencer, so that there are 16 channels on port A, another 16 on port B and a further 16 on port C. If a different 16-part multitimbral synth module is connected to each of these ports then this means that 48 different sound sources

are

available,

each

of

which

can

be

addressed

individually by specifying a MIDI channel and a port letter (A, B or C). Figure

4.3 shows

how

such

a multiport

system

might be configured. It’s important to realise, however, that your multiport interface must be supported by the sequencing software you choose, and lists of suitable interfaces are usually displayed in sequencer manuals, but consult your dealer if you’re in any doubt at all. Hardwarebased sequencers tend to have only one or two output ports, and with no means for future expansion this constitutes one of their greatest limitations in a complex MIDI setup.

Audio out to mastering machinemonitors

Soundcard output

A Soundcard input

MIDI modula wilh stereo outout ak

MID) Outs

IBC

MIC! Ip Multi-port MIDI interaco =|

| iehPaEERe

~ Computer running audio oe JIL

Local On

ncer software saque 3

Master keyboard Figure 4.3: How a multiport system might be configured

Common Problems A basic MIDI sequencing setup will start at the keyboard - it’s here that the MIDI information originates before it’s recorded. The master keyboard is connected via its MIDI Out to the MIDI In of a MIDI interface or directly to the MIDI In of a hardware sequencer or Atari ST. If your keyboard includes a synth section (if it makes

sounds,

in other words)

then switch it to Local Off and patch a MIDI cable from the sequencer’s MIDI Out to the keyboard’s MIDI In. If you have other MIDI modules connected to the system, these can be daisy-chained in any order by feeding the MIDI Thru of one piece of equipment to the MIDI In of the next. Alternatively,

a Thru box can be used.

If your master keyboard doesn’t have a Local Off facility, consult your sequencer manual to see if you can disable the MIDI Thru on the channel to which your master keyboard is set. Most sequencers provide for this. Up to three modules can normally be chained in this Out-toThru setup without problems, but longer chains can cause notes to be stuck or missed (due to the MIDI signal becoming corrupted), in which case you should use a multiple-output MIDI Thru box on the output of your sequencer and then feed each module (or a short chain of two or three modules) from separate outputs on the Thru box.

MIDI Timing MIDI has a measurable timing resolution because the computer’s internal timing routine is based on an electronic clock.

However,

MIDI

is still much

more

accurate

than

a

human performer and is capable of resolving a 4/4 bar of music into at least 960 time divisions. Some software extends this to an even finer degree of resolution, but there are still situations that can cause timing problems. MIDI

is a serial data protocol,

moves

in single file. Because

which the data

means

that all data

speed

is reasonably

high this isn’t usually a problem, but if too much data is sent at once - for example, if all 16 channels try to play a large musical chord at exactly the same time - this results in a musical traffic jam, causing the notes to be spread out slightly. Using a lot of other MIDI control data can also slow things up, but the better sequencing software packages reduce this problem by giving priority to note timing. In most musical compositions MIDI timing shouldn’t be an

issue, but if problems do arise one tip is to put the most time-

sensitive parts, such as drums, onto the lowest-numbered MIDI channels and less critical parts, such as slow-attack strings, onto higher-numbered tracks, in which case the lower channel numbers will be dealt with first. Also, if you need to have a lot of parts playing at once, a modern multiport MIDI interface with intelligent buffering, such as Emagic’s AMT series, will tighten up the timing noticeably. All but the most basic sequencers provide some facility for automating features such as MIDI level and pan, so that sounds can be programmed to change in level or move from one side to another automatically every time the track is played back. This automation works by providing easy access to MIDI controller information, often in the form of graphic displays which can be moved to create fades, level changes and so on.

Sequencers With Audio Using a sequencer with the MIDI-only version limited only by the type quality of the signal

audio isn’t very different from using and the quality of the recording is of audio soundcard you have and the fed into it. On all of the major

sequencers, the audio tracks can be mixed down into stereo

within the sequencer, with full control over channel level and pan at the very least. When a track is recorded it’s actually saved as a file on the hard drive, but it also shows up on the sequencer Arrange page. Here the part can be dragged to new time locations or even dragged onto a different audio track. It’s normally possible to divide the audio parts into smaller sections and move these parts around independently, although it’s also important to note that none of these functions actually change the audio file that has been recorded onto the hard

drive.

Depending on the sophistication of the audio package, you may also be able to edit the audio down to the waveform level, though be warned that most edits at this level are destructive

(ie

they

affect

the

original

file).

Those

edits

routinely available include the ability to change audio levels, reverse files (so that they play backwards), create fade-ins and-outs and replace unwanted regions with silence. Further functionality may also be provided, depending on the sophistication of the software package; for example, there may be functions which can independently change the pitch or duration ofa file. Once the audio material has been recorded and moved into place, the mixing side of the program will invariably allow the tracks to be balanced and panned. Most serious sequencers also provide MIDI-controlled audio automation so that operations such as the movement of level and pan control can be recorded and stored. This will allow you to fine tune your mix until everything is perfect, and then, every time it’s played back, it will have the right levels and the right moves. The number of tracks that can be played back at any time will depend both on the speed of your hard drive and on the power of the computer. Unlike tape machines, on a sequencer it’s usually possible to record as many tracks as is desirable, as long as you don’t try to play back more than the system will allow at any one time. You could record a dozen or so takes of a solo and then choose the best one from this group, for example, and as long as you only play one version at a time it only counts as one track. All you need do is mute the tracks you’re not using. The ability to store alternative takes is often known as having virtual tracks.

The offer also such

mixing capabilities of a MIDI-plus-audio sequencer will level and pan control at the very least, but they will often offer EQ. Unless you have specialised hardware, as a card system with its own DSP (Digital Signal

Processing) facilities, EQ is implemented with the host CPU’s

(Central Processing Unit’s) own processing power. In practice

this may mean that, unless you have a very powerful computer, the amount of EQ you can use is limited. Similarly, there may also be computer-driven effects, such as reverb or delay, which also obtain their power from the computer. Because not all manufacturers of sequencers specialise

in

audio

processing,

more

advanced

processing

functions are usually available in the form of plug-ins.

Software Plug-ins Once the sole domain of specialised hardware, many of the traditional recording and signal-processing tasks are now available in software form, either running on the native processing power of Macs and PCs or on DSP cards optimised for audio applications. For the more sophisticated user, who may need to run a number of powerful processing and mixing applications at the same time, a typical desktop computer may have insufficient processing power, in which case the use of a DSP-based audio interface system is recommended. However, modern computers should be able to run several plug-ins and many channels of EQ without running out of steam.

Third-Party Plug-ins Plug-in effects and processing software is available from a number of third-party suppliers, which means that the functionality of the host software may be extended in various ways. Different versions may be required depending on whether the host system runs on a Mac or PC, however, and

on whether the effects are to run on the native processor or on a DSP

system.

DSP

chips

are not generally

compatible

with those of a different design, and new generations of DSP are appearing all the time, so it’s important to make sure that you buy the correct version of plug-in for your system. For example, Digidesign’s DSP-based Pro Tools hardware will run their own TDM plug-in format but won’t run Steinberg’s VST. At the time of writing, the most common

standard for real-

time plug-ins using the computer’s native processing power is

VST (Virtual Studio Technology). VST was created by Steinberg, the manufacturers of Cubase, but was then opened up to all of the major music software companies in order to promote compatibility. VST plug-ins are available for both Mac and PC platforms, and are supported by a large number of third-party software writers. The standard VST II plug-in expands on the original concept by allowing plug-ins to read MIDI information from the host sequencer. For example, a pitch-shifting plug-in may be able to read MIDI notes directly from a sequencer track, while a delay or modulation plug-in may be able to link its activities to the track’s tempo. Indeed,

because a VST II plug-in can read MIDI note data it’s possible to build a VST plug-in synth - and yes, Steinberg have done that too!

Native Plug-ins Native plug-ins rely entirely on the host computer’s processor and memory for their operation, and the amount of plug-ins that can be run at the same time will depend on the size of the processor. Some plug-ins require more processing power than others, and a good reverb package can use up a significant proportion of your processing resources. Sophisticated de-noising and de-clicking software is also processor intensive.

One way around the need for a powerful processor is to use a

non-real-time plug-in such as AudioSuite or Premier. These generally allow the user to audition a short loop of audio held in RAM while adjusting the various parameters, after which the entire audio file or section of file is processed off line (which is, in effect, a destructive edit). Depending on the type of software package used with the plug-in, it will be deployed in different ways - real-time effects may be patched in via virtual channel or buss insert points, or aux sends and returns, although, on stereo editing packages which have no mixing facility, the plug-ins may be inserted in line with the audio output.

Multiple Plug-ins One of the advantages of using plug-in effects and processors is that there’s no wiring to worry about-they don’t take up any space in your studio and don’t require a patchbay to be connected into the system. Furthermore, all settings are saved along with song data, so it’s not necessary to remember which effects were used or how much of each was added to each channel. Some DSP-based systems (such as Digidesign’s widely supported TDM platform) allow a single plug-in to provide multiple occurrences of the same processor, each with different settings. For example, you may be able to load in a compressor plug-in and then use six of those compressors in different channels of the virtual mixer, all optimised for different programme material. The practical limit on how many occurrences of a plug-in may be used at the same time is the amount of DSP processing power and RAM available, although some plug-in manufacturers may also build a limit

into their software. More sophisticated systems allow additional DSP cards to be installed when more processing capability is needed.

Types Of Plug-in Virtually any effect or form can be built into software equivalents equalisers, gates and general-purpose EQs, emulations of specific

processor that can exist in hardware a plug-in - indeed, there are direct of famous-name compressors, reverbs, as well as any number of delays and so on. Although digital analogue processors are not always

entirely accurate, there are several digital processes that can

be carried out more effectively in a software format, not least because of the more sophisticated displaying and interfacing possibilities accorded by the computer.

In addition to the more obvious processes, plug-ins can also handle de-noising, de-clicking and other high-level tasks, such as azimuth correction on analogue tape masters made on improperly aligned machines. Plug-ins are also available which provide physical-modelling guitar amp simulators, elaborate dithering systems for optimising bit reduction (used when converting 20-bit recordings to 16-bit CD masters), vocoder imitation, and even the deliberate addition of noise

and crackle to simulate vinyl recordings. These latter devices are fashionable for making new recordings sound like vintage

vinyl samples, and include such parameters as record speed, recording age, amount of surface damage and other criteria. Processes such as surround sound, mixing and encoding are also available through the use of plug-ins, enabling a multichannel audio workstation to handle sophisticated TV and film sound mixing. The plug-in environment also allows a few more off-the-wall ideas

to

flourish,

such

as

multiband

fuzz

boxes,

unusual

dynamic equalisers with user-adjustable compression curves and

frequency

points,

and

various

interesting

metering

systems for viewing stereo image, phase, frequency content and so on. There are also numerous pitch-manipulation

packages, including those designed to tune up vocal tracks, stereo width enhancers, 3D sound-positioning algorithms, many types of spectral enhancer and even systems that allow the user to analyse the spectral content of commercial recordings and then automatically EQ recordings to the same spectrum. At the time of writing the majority of plug-ins are designed to be used within software-recording and-editing environments, but manufacturers of some digital mixers are already adding DSP to their products so that effects and processes from third-party designers can be created within the mixer. This can be only a good thing, as it provide the user with more choice, and may turn out to be a good way of offsetting the shortage of analogue insert points on budget digital mixing consoles. One clear advantage of using software is that computers are able to provide an excellent graphical interface and so, as well as controls and meters, plug-ins often provide dynamic

graphs of criteria such as compressor slopes and EQ curves, and this doesn’t take up any extra rack space. However, for most

audio

software,

a monitor

of 17

inches

or more

is

recommended.

Plug-in Automation It’s a well-known fact that some of the parameters of MIDI effects may be automated with MIDI controller data, but many plug-ins may now also be automated, including the popular VST range. While some processes don’t lend themselves to automation in any obvious way, effect such as delay, reverb and even guitar-amp simulation can be automated to produce very creative mixes.

Software Protection

Software piracy is a major problem in some sectors of the industry, and so music software tends to be protected, which can be inconvenient for the legitimate user. However, it prevents unlawful copying of the software, which in the long term should mean better product support and lower prices for the user. Protection comes in several forms, the two most

common being hardware keys, or ‘dongles’, and floppy-disklimited installs. Dongles plug in series with a computer port, such as the keyboard buss on a Mac or the serial mouse port on a PC, and without them the software won’t run. The user can make multiple backup copies of the software if required, but the copies will only run on a machine that has the correct dongle plugged

in.

Some

manufacturers

of plug-ins

use

a single

dongle to authorise their plug-ins, and this usually works with some kind of password system: when you’ve paid for the plug-in the manufacturer supplies you with a password that effectively upgrades your dongle, allowing you to run the most recent plug-ins. Copy protection enforced by limited disk installation uses an uncopyable master floppy disk, from which the software is installed. The master disk usually allows the software to be installed twice, and after each install a counter on the floppy is decremented. When there are no more installs left, the only way to persuade the master disk to work again is to de-install the software (again, using the master disk), which increments the counter on the master disk and provides you with one

more install. The downside of this system - other than the risk of damaging your master disk - is that, if you need to change computers or reformat your hard drive, you first have to de-install all copy-protected software, and if you have a lot of software this can be a long job. This system has a limited life, however, as floppy disks are gradually being phased out.

A more

recent and less intrusive system of copy protection

involves using an uncopyable The

software

CD-ROM

is installed from this CD

as the master disk. ROM

in the normal

way, but on random occasions (usually when the program is started

up)

the user is asked

to insert the master

CD-ROM

before work can continue.

Computing Power The concept of having effects that can run in real time on a host computer with no additional hardware other than a simple soundcard is clearly very attractive, but in order to avoid disappointment you'll need to use the most powerful computer you can afford. Furthermore, you must resign yourself to upgrading your computer every year or two in order to keep pace with newer and more power-hungry software developments. The minimum system requirements stated for software can make it sound like you can get away with using a much less powerful machine, but this is invariably a false impression, especially if you want to run plug-ins. As a rule of thumb, any machine that wasn’t state of the art six months ago is to be avoided, as it’s probably too slow.

Stereo Editors Stereo editors are designed to handle a number of tasks, the most important of which are designing sounds for samplers, editing complete stereo mixes to create new arrangements and assembling final mixes in order to create a finished album master. To perform this successfully, the software must offer the ability to edit precisely and introduce variable crossfades between

sections, the ability to create fades, the

ability to change levels and the ability to silence unwanted material. Plug-in support is also important, as you may want to

equalise

or

compress

the

material.

In

sound

design

applications, you may also wish to use plug-ins that allow you to process sounds in more abstract ways. In addition to creating masters for albums, there is also an increasing demand for software that can prepare files for being loaded onto the Internet using compression protocols such as MP3. It’s also fairly commonplace for stereo editing packages to support direct CD-R burning facilities so that, once you’ve mastered your album, you can produce an audio CD directly from it. As long as you have software that offers full ‘orange-book’ coding (ie it arranges all of the track codes and table of contents in a standard way), you should be able to get a commercial mass-produced CD cut from your CD-R master.

Software Synths Synths are becoming increasingly important now that computers are available with adequate CPU power to run them. Computers can handle any type of synthesis, from emulated analogue to FM, sample based, wavetable, additive and even granular. The amount of available polyphony is usually related to the amount of available computer processing power, and such soft synths really come into their own when offering synthesis methods which aren’t easily available with existing hardware. Some driver software is necessary to allow the synths to receive MIDI data and play through your audio hardware, though ASIO support is fairly common. Steinberg’s ReWire technology makes the system even more flexible by enabling the output from a virtual synth to be fed through the mixer in the host audio program, which means that it can then be mixed and effected just like any other audio source. Opcode’s OMS (Open MIDI System) software is also commonly used in Mac systems to route MIDI-to-OMS-compatible pieces of

software. The main problem with software synthesis is that the latency of the audio hardware may be great enough to prevent the synthesiser from being playable. This sounds serious, but in practice the part can often be recorded with a soundcard synth voice and then played back on the soft synth. The adoption of this method avoids the effects of latency.

Samplers Software samplers are now available which operate much like their hardware counterparts, some using system RAM and others streaming audio directly from the hard drive. Their MIDI control and audio outputs are handled in much the same way as soft synths, so you'll need the right drivers in order to use them in your system. This is definitely a growth area, but SoundFont soundcards developed by Creative Labs have had sampling capability for a long time, and the quality on some of their better-specified cards can be impressive. On cheaper cards, sample RAM must be plugged into the card itself, while on more sophisticated systems it’s possible to store sample data on the computer’s own system RAM. In essence, a soundfont is a set of sound samples that can be loaded into a card for playback. Numerous CD-ROMs of soundfonts are commercially available, while the support software that comes with most Soundblaster cards allows .WAV files to be recorded and then edited to create new soundfonts. Although the Soundblaster name is associated with budget games cards, the SoundFont system is actually pretty sophisticated and the quality of some of the commercial

sample

material

available

for it is impressive.

What’s more, SoundFont material tends to be considerably less costly than more professional sample formats.

5 MUSIC COMPUTERS

In the preceding chapters we’ve learned that computers can supply us with MIDI sequencing, audio recording, mixing, virtual

effects,

virtual

synths

and

CD

manufacturing,

but

because of the complexity of the software and the potential for conflicts between software and hardware, setting up a desktop studio isn’t always as simple as it might first appear. In this chapter I’ve put together a few suggestions that will help you to choose the right system, and if you already have one it will help you to use it to its full potential, whether it runs on a Mac or a PC.

If yow’re planning to buy a PC system but aren’t sure which sort to buy, check out the FAQs on the various music web sites, including the extensive FAQ section in the Forum section of the Sound On Sound site at www.sospubs.co.uk. If you know somebody who has a working system and you’re prepared to order exactly the same hardware and software, you should be on reasonably safe ground (although parts occasionally change within different models of the same computer).

However, those with limited computer experience should seriously consider buying a pre-configured system from a single vendor. Choosing a Macintosh isn’t as difficult as choosing a PC because there’s a limited range of models, as opposed to the infinite permutations of PC configurations. All new Macs use the USB serial interface for the connection of peripherals such as MIDI interfaces, and from the tests I’ve done so far

those with intelligent buffering - such as the Emagic AMT series and the MOTU USB interfaces - provide the best MIDI timing. Newer Macs don’t have built-in SCSI facilities, however, so if you want to use the latest and fastest SCSI drives you’ll need to fit a SCSI PCI card, just like a PC. You should always contact the software manufacturer’s technical support department before buying a SCSI card, or a MIDI interface for that matter, just to ensure that there are no known problems with your selection. If you need fewer audio tracks,

a slower

(but

cheaper)

internal

EIDE

drive

may

be

adequate, but I’d still recommend buying a separate drive rather than using your boot drive. Most Macs have space for a second drive, but if you absolutely must use the same drive for your audio work as the one on which you do everything else, at least create a separate partition for your audio work. The newer and faster FireWire drives look to be the future of

Mac and PC audio storage, but I haven’t yet been able to try one.

Note that Macs from the coloured G3 iMacs onwards don’t have floppy drives as standard, though it’s possible to buy an external drive that takes regular floppies as well as highcapacity removable disks. Check with your software vendor that this doesn’t complicate software installation or the copyprotection system in any way.

Recording Levels It’s best to optimise the input signal level of any digital recording system at source rather than relying on normalising your system or using software gain control to bring up the level. In most cases, gain control exercised in the software domain simply boosts the low-level signal at the input and brings up any noise by the same amount. You should use the level metering provided in the software to

keep peak levels just a few decibels below the point at which clipping occurs for the least distortion and best signal-tonoise ratio.

Regardless of whether your system is capable of 16-, 20- or 24-bit recording, the ultimate sound quality will be defined by the recording made at source. In most cases, the dynamic range of a 16-bit system will exceed the dynamic range of the signal being recorded. For vocals, consider buying a voice channel/pre-amp type of device, which is essentially a good mic amp equipped with EQ and compression. This may also be used when miking other instruments, and many such amps feature an instrumental DI input suitable for bass and clean electric or electro-acoustic guitars. Alternatively, you could

use

the

mic

pre-amp

of a mixer

that

you

know

is

suitable for recording.

Weak Links The such you rest

fact that computers and recording software packages are good value for money can lead you into believing that can make do with using equally cheap components in the of the studio, but this simply isn’t true, although

premium

items such as capacitor mics are now much

more

affordable good mic, mixer - a with using instances results.

than they used to be. You should budget for a decent cables and - if you don’t have a suitable mic pre-amp/voice channel. You could get away the old dynamic mic you use at gigs, but in most a capacitor mic will provide noticeably better

This also applies to your monitor speakers. If they’re inaccurate you won’t really know what your recording sounds like. Use high-quality monitor loudspeakers on rigid stands and set them up so that you’re at the apex of a roughly equilateral triangle, with the monitors pointing

directly toward you. You don’t need to monitor loudly, but you'll need enough volume to overcome the noise made by the fans and drives in the computer. If you can put your computers in a ventilated cupboard to reduce the noise then so much the better, and lining the cupboard with an inchthick layer of fireproof furniture foam will also help to damp out noise.

Drive Maintenance If you have a separate hard drive for your audio work you’ll be able to regularly defragment it, or even reformat it, without disturbing your program files. Defragmenting optimises the rate at which data can be read on and from the

drive - a badly fragmented drive can produce audible glitching and reduce the number of tracks that can be played back simultaneously. Faster drivers can play back more tracks, although very fast drives may need a special, rapid SCSI interface card to make the best of their abilities. Modern hard drives are reasonably reliable, but it’s still wise to back up important work to another drive or to another medium, such as CD-ROM. Avoid moving or knocking drives while they are operating, as this can cause serious damage. When choosing or upgrading a soundcard, try to find one that provides at least four outputs, and a digital S/PDIF Out if you own a DAT machine or MiniDisc recorder. In this way you can use one pair of outputs for tracks that use software-

based plug-in effects, while the other outputs can carry tracks that you want to effect with external processors.

Effects Reverb is the most important effect in the studio, and effective reverbs take up a lot of computing capacity. For this reason, it may be worth considering buying a soundcard with its own hardware

reverb processing capabilities, or a multi-

channel audio interface that allows you to use an external hardware reverb unit. Unless you are using a fairly sophisticated soundcard with on-board DSP processing, you’re likely to experience some latency or delay when monitoring the signal you’re recording through the system. ASIO II drivers will minimise this problem in compatible hardware, but it won’t cure the problem in all soundcards. One way around this is to use a small mixer, and to monitor the computer’s input rather than its output when overdubbing. By monitoring the input source you will avoid problems with latency, but you will also have to monitor without using plug-in software effects. You will usually need a separate mixer to combine your audio and external synth and sampler signals, and if you hook up a hardware reverb unit to this mixer you'll be able to add reverb to the monitoring as you record. This makes singers feel more comfortable, and can help them to produce a really good vocal performance. You don’t need to record the reverb while monitoring; if you leave it until mixing before adding it you keep your options open, and you can then try different settings.

Other Plug-ins Not all plug-ins are effects - some are designed to fix problems. The Antares Auto Tune plug-in is particularly useful, not only in cleaning up vocal pitching but also in tightening up guitar solos, and a cheap VST ‘lite’ version is available that will work with most MIDI/audio sequencing software. Regular playing will be unaffected as long as the tracking time is set at a slow enough speed, but whenever a note is sustained it will automatically settle on exactly the right pitch. This can be particularly useful for slow pieces that use a lot of string bends.

Every month newer and more sophisticated software plug-ins become available that handle anything from automatic dialogue replacement (for film and TV work) to mastering. The best way of keeping up to speed with what’s on the market is reading recording magazines and visiting manufacturers’

web

sites.

However,

it’s possible

to get by

with a fairly basic set of plug-ins; for most music work, a good reverb unit and compressor, combined with a delay/echo effect, will satisfy most requirements. Check out the plug-ins that come bundled in with your software before forking out for any new ones, as some of these freebies can work very well.

Guitars And Computers One problem that most guitarists come up against is that computer monitors interferes badly with the guitar pickups, resulting in a buzz being picked up on the recording. Some humbucking pickups are reasonably good at rejecting this buzz, as long as the player doesn’t sit too close to the monitor while recording, but single-coil pickups affect the sound very badly, especially if overdrive is also used. A cheap way around this problem is to switch off the monitor just before recording and use keyboard commands to start and stop the recording process. If for some reason you can’t switch off the monitor, sit as far away from it as possible when recording and rotate your position to find the null point (the point at which the buzz is least obtrusive). You might also use a noise gate pedal to keep your guitar quiet between phrases. Flat-screen LCD monitors are becoming cheaper, and these take up less space

and eliminate the electromagnetic interference generated by the scan coils of CRT monitors. These are good investments if you record a lot of guitar work or you’re short on space.

Fan And Drive Noise Physical noise is also a problem when miking instruments or voices in the same room as the computer. If possible, turn off unnecessary external drives, CD-ROM

burners and so on, as

these often make more noise than the main computer. Set up your mic (ideally a cardioid model) as far from the computer as possible, and improvise an acoustic screen between the mic and the computer with a duvet or sleeping bag. Also, make sure that the mic is pointing toward an absorbent rather than a reflective surface. Work as close to the mic as possible without compromising the sound, always using a pop shield for vocals. This will be easier if you use a sounddeadening cupboard, as described earlier in this chapter.

Waveform Editing Virtually all sequencers capable of recording audio have a Waveform Edit page (although it isn’t always called this), where it’s possible to highlight and silence selected portions of audio. If background noise is a problem, matters can sometimes be improved by manually silencing the gaps between words and phrases, which doesn’t take as long as you think and can really improve the quality of a recording, especially when there are multiple audio tracks. It’s a good idea to normalise audio recordings before processing them in order to maximise playback signal-to-noise ratio, although you shouldn’t use this as a substitute for arriving at the right recording levels in the first place. Normalising can generally be carried out from within the Waveform Edit page. Waveform editing can also be used to remove unwanted finger noise from between notes in guitar tracks for those occasions when you end up with a take that’s almost perfect, except there’s too much squeak or finger noise between notes, or maybe the next string was caught just after a note bend. The silence function will surgically remove these little

errors, although if you leave them where they are, and instead reduce them in level by between 6dB and 20dB, you may end up with a more natural sound. In this way you can also silence page turns, coughs and other noises on vocal parts. You can also be very creative, such as adding fade-ins to guitar parts to simulate the action of a swell pedal.

For the greatest possible flexibility when adding effects in the mix, try to record all parts dry - in particular, don’t add reverb or delay unless you really have to. Add reverb at the monitoring stage if you need to hear it to create a good performance, but don’t record it. In this way you'll be able to edit tracks without cutting holes in the echo or delay effects you’ve added. Furthermore, if there are any edits that are a bit close for comfort, adding the necessary delay or echo afterwards will help disguise the edit points, making the recording sound quite natural.

Computer Power As with any other software, plug-ins always take up a certain amount of the available computing power, so if you want to add the same delay or reverb-based effects to several tracks it’s more efficient to use a single plug-in configured as an aux send processor than to use a separate insert plug-in on every track. The aux send controls can generally be used to add different amounts of the same effect to any track, in the same way as a regular mixer, all for the CPU overhead of a single plug-in. You should note, however, that under normal circumstances you can’t use the aux send with processes such

as EQ, compression or gating - these have to be used with inserts, though you may be able to premix to a group buss and carry out one lot of processing to the sub-mixed tracks rather than having to process each track individually. Don’t try to force your software to do everything for you just

because it can; very often you'll find that you can get a better sound with separate devices, and of course they won’t burden the processing power of your CPU. You can still compress signals as you record them, even if you don’t have a multioutput soundcard, ideally by using a voice-channel-type device as described earlier. This also applies to EQ. Only the best digital EQs sound as natural as even the most basic analogue equalisers.

Time And Pitch Manipulation There are lots of tricks you can perform with the audio manipulation facilities provided by your sequencer, and many of these extend beyond what you might expect from a simple multi-effects unit. For example, there are very powerful filters that can process any sound to make it appear synthetic, and devices which can distort sound in other interesting ways. Built-in facilities vary from one software package to another, but pitch-changing and time-stretching are supported by most machines. These are invaluable for massaging audio sample loops, but you may also find other tools which provide level-maximising, de-noising and other features.

Some

of the

more

processor-intensive

processes

work off line, so they can be used even on slower machines you just have to wait a little longer for the results. Pitch-to-MIDI conversion is also fairly common. This process allows a monophonic recording of an instrument to be analysed in order that a MIDI track mirroring the note and pitch bend information of the original track can be produced. This is useful when you need to configure a MIDI instrument to play along in unison with material from a real instrument

that you’ve just recorded. Some software ability to quantise audio tracks, which means of parts such as drums and guitar riffs can considerably. In my experience this can work

even offers the that the timing be tightened up well, as long as

you don’t process too much material at once and as long as the original audio information isn’t too complex. As with so many things, you’ll need to experiment to see what works and what doesn’t.

Keep It Clean! Most computer audio systems run best if you get rid of any superfluous software, such as screen savers and games. You should also make sure that you have no more drivers than you actually need (or extensions, if you’re using a Mac) - the cleaner your system the less likely you’ll run into problems. Also, check manufacturers’ web sites to make sure that you have the most up-to-date drivers or extensions, as improvements are being made all the time. If you must run other software, consider using the dual-boot system described in Chapter 3, with which two completely different sets of the operating systems, along with the relevant programs, are arranged on two different hard drives. In this way you can start up from your music drive when you’re working with audio and from the other drive for everyday work or for playing games. It’s also wise to carry out tests to find out how many tracks and plug-ins your machine can run without running into trouble - try to work with no more than half to two thirds of this amount. Most sequencers are equipped with some kind of CPU activity monitor, but the demands on your CPU aren’t constant. Disk drives also slow down as they fragment, so try to allow for this - although you can’t be expected to defragment it after recording each track.

Computer Timing Problems occasionally arise where the timing of MIDI and/or audio isn’t as accurate as it should be. MIDI timing errors (caused by the speed of MIDI data transfer) are small

compared to the erratic timing of even a good player, but if you have a composition comprising many MIDI tracks, all of which contain quantised notes at the start of a particular bar, the errors may become noticeable. Fortunately, there are strategies that can be used to improve matters. You should switch off any type of MIDI message you’re not using, so if the track you’re recording doesn’t require aftertouch it should be switched off at the keyboard. Similarly, if you’re using a MIDI guitar controller and the track doesn’t require any pitch bend, filter it out or switch to Chromatic mode. Having economised on your data stream in this way, there are also other tricks you can try. All MIDI instruments take a short time to respond to incoming MIDI data, which can exaggerate any delays that occur normally. If you have an instrument that you know is relatively slow, try applying a negative delay to the track in your sequencer’s playback parameter section so that the MIDI data arrives a few milliseconds before the beat. This will speed up the response of the synth and also avoid the bottleneck of too many notes trying to squeeze out on one quantise segment. A similar trick can be achieved with slow attack sounds, such

as strings - it’s possible to advance most slow string or pad sounds by several milliseconds before any timing differences can be heard. Indeed, it’s always best to avoid having too many quantised notes falling at the same time, so if you can leave something unquantised then so much the better. Back in the old days, when the Atari ST was state-of-the-art technology, it was said that you should put drums and other parts with critical timing on the lowest-numbered tracks, as MIDI channels with low numbers had timing priority. Today’s sequencers are able to move tracks, but it’s still likely that data will be processed in order of MIDI channel, so it’s a good idea to put all parts for which timing is critical on the

first few channels. If none of this fixes your problems, and you’re using a MIDI interface that’s part of a cheap soundcard, you should consider buying a more sophisticated MIDI interface. Before spending any more money, however, see if you can obtain more recent drivers from your card manufacturer, as these can sometimes make a big difference to a performance. Using a multiport MIDI interface may also tighten up timing, although much depends on the speed at which the computer communicates with the interface. If you’re using multitimbral synths with a multiport interface, try to reserve a separate port for each one, even if you’re not using all 16 channels. The newer Emagic and MOTU interfacing systems tighten up timing even further by sending data to the MIDI interface a short while before it is needed, where it’s held in a memory buffer. The data is then clocked out of all ports simultaneously at the appropriate time, providing the best resolution MIDI has to offer.

Lightening The Load Your MIDI and audio timing can be optimised by ensuring that your computer isn’t carrying out any unnecessary tasks

in the background. Scrolling screen displays usually take up more processing power than those that jump when the cursor

gets to the end of the screen, so try turning off the scrolling function if you think you’re pushing your computer’s capabilities. Even better, reduce the magnification of the song’s display so that it all fits into a single page before you play it back - in this way there’s no scrolling or screen redraw. Mac users should also try using the Extensions Manager control panel to disable any system extensions that aren’t

necessary, and screen savers are definitely something to avoid. Configuring the monitor resolution to 256 colours also reduces the processing burden.

Audio Timing Most people can get by with the resolution of MIDI, but if audio material refuses to stay in time with the MIDI tracks it can be extremely frustrating. There are several reasons this can happen. Latency is something that many users have to accept when monitoring while recording, but if it’s too bad it can be improved if the input signal is monitored over a hardware mixer. When the audio material is transferred onto your computer’s hard drive, however, playback timing can still wander around to an unacceptable degree if your computer is short of RAM, if it’s slower than at least twice what the manufacturers claim as a minimum specification, or

if the software drivers are not up to date. This latter consideration is important, because a poorly written driver may allow the playback latency to waver, which has the effect of causing the audio timing to drift. You should also check out the pages of driver settings, adjusting the amount of RAM buffering and playback delay settings until your system behaves properly. Fortunately, many soundcard manufacturers recommended settings to ensure compatibility with the more common software packages. Check that you’re not asking the computer to start playing the audio material on the first beat of a sequence. When the Start button is pressed on a sequencer a certain amount of behind-the-scenes activity (buffer filling, etc) takes place before it’s possible for the audio to start playing back. If the audio material starts on the first beat of the song, give the sequencer time to prepare for playback by inserting a couple of empty bars between the time when the count-in stops and the song starts.

Audio And Clocks When audio material is recorded onto a computer, the data is usually timed by a crystal clock fitted to the soundcard; however, although crystal clocks are very stable, they don’t all run at exactly the same speed as each other, which means that, if you have a system with two or more soundcards, you have to be careful when recording with one soundcard and playing back with another, as the playback speed on both cards may be slightly different. This situation should be avoided if each track is replayed with the card that recorded it, but things can still go wrong in a system if you try to sync all of your audio to an external digital timing. Again, the playback speed may drift slightly if the external clock doesn’t run at exactly the same speed as the clocks on your soundcards, and on long audio tracks the drift becomes more pronounced the further the track progresses. If you intend to synchronise all of your cards on playback, sync’ing them while you’re recording as well will lessen the chances of drift.

Sync What happens when your audio is stable but you decide to sync the sequencer to a less than perfectly stable outside source, such as a tape recorder? If the master device runs a little slow the MIDI sequence tempo will drop to match it, but the audio may then carry on playing at its original rate, causing the MIDI and audio material to drift apart. The better sequencers have something called continuous audio resync, which changes the audio clock speed to match that of the time code produced by the master machine, but not all machines are equipped with this facility. If you’re using a number of soundcards in order to gain more individual audio outputs, it’s also unwise to split stereo pairs of signals between two cards because even the slightest difference in timing will cause an unpleasant flanging effect.

Wherever possible, record stereo material as a stereo audio file - in this way the two halves can’t drift.

Solutions If the stability of your audio timing is still inadequate, despite your best endeavours, one of the most effective solutions to this particular problem is to divide long audio tracks into sections of just a few bars long. Each section will trigger independently so that, although there may still be a small amount of drift within each section, the effect won’t be

cumulative. Cubase VST provides a priority setting in its System Setup page, which defaults to ‘normal’, giving equal priority to both audio and MIDI. If you work mainly with audio, however, you can change this to ‘high’ or ‘highest’, giving priority to audio data processing. If you work mainly with MIDI, on the other hand, setting this to ‘low’ will ensure that MIDI has priority. With VST, it’s also important to have the correct setting for system pre-roll (located in the Sync menu). Normally this should be set to the same value as the soundcard’s latency. When you're creating audio loops within a sequencer, much depends on how slick the sequencer is at retriggering the same piece of audio. If slight timing discrepancies arise when the sequencer jumps back to the start of the audio segment, you may be able to make a small tempo change of usually a very small fraction of a beat per minute to compensate. If you can’t find a satisfactory solution, however, you should perhaps consider entrusting your drum loops to an external sampler. Ultimately, a great deal depends on the quality of the computer. It should obviously be capable of processing MIDI

and audio, but we also expect them to be able to deal with mixing, EQ, VST real-time effects and even virtual synthesis.

Unless you have the most powerful computer, you may find that many problems are simply caused by having an underpowered system. Manufacturers also confuse the issue by quoting hopelessly inadequate minimum specifications on software packages. To get the best out of your system, make sure that you have plenty of memory available - ideally double that of the minimum specified, or more - and take some time to check manufacturers’ web sites to download the latest drivers and setup information for your hardware. If support helplines can’t help you, user groups are probably the best place to start looking for answers to any problems you still can’t resolve.

6 CREATING AN ALBUM

You’ve recorded all of your MIDI tracks, recorded your audio tracks and you’ve set up a mix that sounds great. So now what? Ultimately you need to end up with a stereo master recording, and how you proceed from this point depends on the type of system you’re using. If you have an external DAT or MiniDisc recorder you can mix to this, either straight from your soundcard output or with an external mixer. The external mixer provides more flexibility if used in conjunction with a multi-output soundcard or if you have hardware synths, but if you’re working with a soundcard that provides a single stereo mix of audio and MIDI tracks then there’s little advantage to be had in using a mixer. In fact, best results can be obtained by feeding the digital output from your soundcard (if you have one) directly into your digital recorder’s digital input. Once your mix is on DAT or MiniDisc you will then be able to feed it back into the computer when you want to compile your album. If you don’t have a digital recorder, you'll need to create your master mix as a stereo sound file on the computer and then work on this with your editing software. Different soundcards may handle this procedure in slightly different ways, so refer to your soundcard manual to find out the best way to do this. If you have recorded the original material with a bit depth of more than 16 bits, it should be rerecorded at 16-bit resolution at this stage.

Mastering Once you have a collection of mixed songs stored as stereo

sound files on your hard drive, you may want to put them together to make an album, and with the cheapness of CD-R burners and CD-R discs, why not? However, before burning a CD there may be a few mastering tricks you can employ to make the tracks sound even better. Moreover, you may need to tweak tracks that were recorded or mixed at different times so that they sit together comfortably, and this may require both level and tonal adjustment. Although professional mastering engineers use a lot of esoteric equipment to achieve this, there’s no reason why you shouldn’t get impressive results with software plug-ins. The most important tool is the ear of the person doing the job. I'd recommend a a

good

plug-in tool kit comprising, as a minimum,

parametric

equaliser

and

a

compressor/limiter.

Enhancer-type plug-ins can also be used as de-noising devices if your finished mix has more hiss than you’d like. You'll also need to have an accurate monitoring environment, which means using hi-fi speakers rather than plastic computer speakers with a bass speaker the size of a shoebox stuck under the desk. Most mistakes are the result of over-processing, and the old adage ‘If it ain’t broke, don’t fix it’ holds true with mastering. Don’t feel that you have to process music just because you can, or you might find that your mastered sounds are worse than your original material. Here are a few tips to help you make the best of your recordings. The endings of fade-outs should be handled with the computer editor. Not only does the computer provide more control but it will also fade out any background noise along with the music so that the song ends in perfect silence. You should also silence any background noise that occurs just before the recording so that your track also starts from

complete silence. Endings should be faded out rather than silenced in order to mimic the natural decay exhibited by most instruments. Start to fade when the last note or beat has decayed to around 5% of its maximum level, and make it around one second long so that the song finishes in silence. Including around half a second of silence at the start of each song will mean that slow CD players won’t miss the start of the song. Once you’ve decided on a running order for the tracks you'll need to match up levels, but this doesn’t mean simply making everything the same level. If you do this, more restrained material, such as ballads, will seem very loud when compared to stronger songs. The vocal level will often be the best guide to how well songs are matched, but your ears are the best judges. Close your eyes and try to hear if the songs sit comfortably together, tweaking levels until they sound as natural as possible.

Mastering EQ If the tracks were recorded at different times or in different studios, changing levels may not be enough. A little EQ will often improve matters, although a good parametric EQ will be required to avoid making matters worse. Listen to the bass ends of all songs to determine how they differ, and use EQ to try and even things out. For example, one song might have all of its bass energy bunched up at around 80Hz or 90Hz, while another might have an extended deep bass dropping right down to 40Hz or below, and by rolling off the sub-bass and peaking up the 80Hz area slightly you may be able to focus the bass end. Also, the track with the bunched-up bass could be treated by adding a gentle boost at 40Hz, combined with a little cut at around 120Hz. All equalisers behave differently, so you’ll need to experiment.

To make a track sound louder when it’s already peaking close to digital full scale, use a digital limiter such as the excellent Waves L1 or Logic Audio’s Energizer. The overall level can usually be increased by at least 6dB before the peaks sound processed.

Gentle Compression Overall compression can also add energy to a mix and help to even out the material, but it isn’t mandatory: music needs some light and shade in order to provide dynamics. A compressor will often slightly change the apparent balance of a mix, sO you may need to use it in combination with EQ. If the EQ is applied before the compressor then the boosted frequencies will be compressed the most, while applying it after the compressor will allow you to equalise the compressed sound without affecting the operation of the compressor. The best arrangement to choose will depends on the material you’re treating, so you should try both. A split band compressor or dynamic equaliser will provide you with more scope to change the spectral balance of a mix, but it takes a little practice to use these devices so that you feel like you’re controlling them, rather than vice versa. I find that the best results are achieved by using compression ratios below 1.5:1,

and with the threshold

set at between

-—10dB

and —

30dB. After applying compression, listen to the finished master all the way through over a pair of headphones, which will show up small glitches and noises that loudspeakers may mask digital clicks can appear in even the best systems.

De-noising Digital de-noising programs can’t work miracles - even the best systems produce side-effects if pushed too far. The simpler systems are effectively multiband expanders on

which the threshold of each band is determined after analysing a section of noise from between tracks. It’s therefore best not to clean up original masters before they’re edited, or there may be no noise samples left from which to work. You should be able to achieve a few decibels of noise reduction before the side-effects kick in - as low-level signals open and close the expanders in the various bands, the background noise is modulated in a way that can only be described as chirping. This chirping become worse the more noise reduction you try to achieve, so it’s best to use as little as possible.

Track Editing If you need to make edits within individual tracks when compiling a version from all of the best sections of several mixes or recordings, try to make joins just before or just after a drum beat so that any discontinuities are masked by the beat. However, if you have to use a crossfade edit to smoothe over a transition, try to avoid including a drumbeat in the crossfade zone or you may hear a phasing effect when the beats overlap. Crossfades should be as short as possible in order to avoid producing a double-tracked effect during the fade zone - as little as 10-30ms is enough to avoid producing a click. Once edited, resave the track as a new sound file.

Making A CD Most CD burners come bundled with software suitable for making audio CDs, but if you wish to use the CD as the master for pressing commercial CDs then you will need to choose

software

that offers

‘red-book’

PQ

(Pause

and

Cue)

coding so that the disc you burn has the same data format as a commercial CD. The disc must be written in ‘disc at once’ mode, rather than a track at a time, in order to avoid

introducing

errors

between

tracks.

Check

with

your

CD

manufacturer to confirm that their device can work from CDR as a master, and take note of any special requirements it

may have. Virtually all CD-burning software includes a

playlist feature,

on which individual song files (or sometimes regions of files)

are added into the playlist, and the gap between the tracks can then be determined by the user. The default value is generally around two seconds, but this can be changed if required. Figure 6.1 shows the playlist from Adaptec’s Jam program for the Macintosh, which is pretty typical for such software. When

youw’re deciding

on the space between

tracks on an

album, listen to how the first track ends and how the second

one starts. Gaps are rarely shorter than two seconds, but if the starts and ends of tracks are very abrupt you may need to leave up to four seconds between tracks. Use the pre-roll feature of your digital editor to listen to the transition so that you can judge when the next track should start. a

dC Disc

k

1

a

F,

ES

bf b

7

Goo

2 1E

bo

GORD

Be

b

9

b

10

Reape

42 Bevined aM re Pg

aed dem

eats

fe

fewer

(Oo

BSE

fe

Mew Bag Fine

oto

Teese

fe

Fleck Time Lave Pied

b 12 oc n5:00

42 40 H

boo

ioe

ite

Ole ken

Fined

i} Cetetrate reed

7

Oss7a4

on

gD

UE

-2D

gO

us rea

-af

go

igre

-10

O73

|

an 55:

'

a

OD 0 P

eo SE

- 6

FS

opor oo

aot

le

eerghedy Eeowe Vied 4 0T2Z

OD

gO

b

Ste

me is ao

EF

Regen 14

ao

aC

ta

bos

Goon

thse

BF

ego

18

Boh

OR

Aes

A

Begin

1d

Ti eerie

oF |

bine

nes

Figure 6.1: CD burning playlist

Tees

Of f2e ess Gar |

afer is! ft

|

oo = -3o

I

go ZOD

Lael OT Brakes Bb 27:

7

When you’re ready to burn the CD, load in a blank CD-R, and avoid touching the flat surface because dirt or fingerprints will cause errors in the writing process. You should also resist the temptation to use cheap, unbranded discs, as some of these are prone to high rates of error, short storage lives or both. It’s also a good idea to run the software in Test mode to check that the burn will work, without introducing errors, at

the write speed you’ve chosen. If an error creeps in during a real burn you’ll end up with an expensive coaster!

The Internet Making your audio files available on an Internet site good way of publicising your music, and there are two ways of doing this. In both cases the amount of audio needs to be reduced, because straight .WAV files are large and take forever to download.

is a main data very

One approach is to use

the popular MP3 data-compression format, which reduces audio files so that they are less than a tenth of the size of the original .WAV file and yet retain almost all of the original subjective audio quality. MP3 files can be downloaded like any other file and then either played back by the computer or a hardware MP3 player. MP3 is the most popular format for selling audio material over the Internet, but even so one minute of stereo audio will still require something like 1Mb of data. The other method is to use a lo-fi streaming format, such as RealAudio. These files are much more heavily compressed than MP3, and sound noticeably compromised, but they’re still of sufficient quality to allow listeners to tell whether they like the music or not. The advantage of using a streaming file is that, once a small part of it has been downloaded, that part of the file can be played as the rest downloads - you don’t have to wait for the whole thing to be downloaded. This makes it suitable for those casually

browsing record catalogues or musicians’ web sites. You'll need encoding software to turn .WAV or SDII files into MP3 files or a streaming format like RealAudio (although MP3 encoding is now included in some editing packages), but the recipient can usually download a free program from

the software vendor to play them. You should therefore always include a link on your site so that visitors can obtain this software. MP3 playback is often supported via plug-ins for existing browsers.

APPENDIX

Common

2

ee

Cable Connections

eee

Unbalanced jack to jack

Tinh

6 To

Soouc &

Oo Reeve

eee Balanced pack 14 jack

lin

oT

Hing & CCC

Balanced la unbalanced jack

© Ang

a

"Ir o Pry —T_5 Bb

$a =A a Screen left unconnected at

es

© Tip rrr

unbaknced and

TAS 10 mone jack ¥-lead nee polnts)

Tic a Any

Secve

© Tp Sanva

To Bee

TRS 10 XLA Y-lead

(nga paints) Nor all INS

PONS

Ae wired 1

ihe same convention. so

check your mixer manual

“Ip &

—® 1 Sarecr

2-4

Arg

ee

LG)

He 7 Scroor

Cit

ee

2 cd

9 Codd

Unbahinced 4LF 10 XLA

;

ce

5 4

-

01 Siar

Qf —c4 ea 3 ated

Balanced XLA to XLR ie

a do

1 Sheer

© oon Codd

Screen lett unconnected at unbalanced and

eS

pin 3 when using Unbalanced

Consult equpmen! details



, =

& Socp4

manual for

Urialanoed jack ta balanced XLR

__. . “if

Rirg Segae oe

—i

1

ei 2 +]

n

ih

Note: on some equipment, XLA pla 1 must be joined 10

$6

Balanced XLA to unbalanced jack

GLOSSARY

Active Describes

a circuit that contains

transistors,

ICs, tubes and

other devices that require power to operate and are capable of amplification.

A/D

Converter

Circuit for converting analogue waveforms into a series of values represented by binary numbers. The more bits a converter has the greater the resolution of the sampling process. Current effects units are generally 16 bits or more, with the better models being either 20- or 24-bit.

AFL After-Fade Listen, a system used within mixing consoles to allow specific signals to be monitored at the level set by their fader or level control knob. Aux sends are generally monitored AFL rather than PFL so that the actual signal being fed to an effects unit can be monitored.

After touch Means of generating a control signal based on how pressure is applied to the keys of a MIDI keyboard.

much

Aliasing When an analogue signal is sampled for conversion into a digital data stream, the sampling frequency must be at least twice that of the highest frequency component of the input signal. If this rule is disobeyed, the sampling process becomes ambiguous, as there are insufficient points to define each

waveform cycle, resulting in enharmonic sum and difference frequencies being added to the audible signal.

Ambience The result of sound reflections in a confined space being added to the original sound. Ambience may also be created electronically by some digital reverb units.

Amplifier Device that increases the level of an electrical signal.

Amplitude Another word for level. Refers to sound or electrical signal.

Analogue Circuitry that uses a continually changing voltage or current

to represent a signal. The term originates from the fact that the electrical signal is analogous to the original signal.

Anti-Aliasing Filter Filter used to limit the frequency range of an analogue signal prior to A/D conversion so that the maximum frequency does not exceed half the sampling rate.

Attack The

time

it

takes

for

a

sound

to

achieve

its

maximum

amplitude.

Aux Control on a mixing console designed to route a portion of the channel signal to effects or cue mix outputs.

Aux Return Mixer inputs used to add effects to the mix.

Aux Send Physical output from a mixer aux send buss.

Balance This word has several meanings in recording. It may refer to the relative levels of the left and right channels of a stereo recording, or it may be used to describe the relative levels of the various instruments and voices within a mix.

Balanced Wiring Wiring system which uses two out-of-phase conductors and a common

screen to reduce the effect of interference.

Bandwidth Means of specifying the range of frequencies passed by an electronic circuit such as an amplifier, mixer or filter. The frequency range is usually measured at the points where the level drops by 3dB relative to the maximum.

Boost/Cut Control Single control which allows the range of frequencies passing through a filter to be either amplified or attenuated. The centre position is usually the ‘flat’ or ‘no effect’ position.

Bouncing Process of mixing two or more recorded tracks together and re-recording these onto another track.

Buffer Circuit which isolates the output of a source device from loading effects due to the input impedance of the destination device.

Buffer Memory

Temporary

RAM

used

in

some

computer

operations,

sometimes to prevent a break in the data stream when the computer is interrupted to perform another task.

Buss Common electrical signal travel. In a mixer, there are mix, the groups, the PFL Power supplies are also fed

path along which signals may several busses carrying the stereo signal, the aux sends and so on. along busses.

Cardioid Literally ‘heart-shaped’. Describes unidirectional microphone.

the polar response

of a

Channel In the context of MIDI, the term channel refers to one of 16

possible data channels over which MIDI data may be sent.

Channel In the context of mixing consoles, a channel is a single strip of controls relating to one input.

Chorus Effect created by doubling pitch modulation.

a signal and adding

delay and

Chromatic Describes a scale of pitches rising in steps of one semitone.

Clipping Severe form of distortion which occurs when a signal attempts to exceed the maximum level which the equipment can handle.

Compressor Device which reduces the dynamic range of audio signals by reducing the level of high signals or by increasing the level of low signals.

Conductor Material that provides low resistance for electrical current.

Console Alternative term for mixer.

Cut-And-Paste Editing Copying or moving parts of a recording to different locations.

Cutoff Frequency Frequency

above

or below

which

attenuation

begins

in a

filter circuit.

Cycle One

complete vibration of a sound

source or its electrical

equivalent. One cycle per second is expressed as 1Hz (Hertz).

Daisy Chain Term

used

to

describe

serial

electrical

connection

between

devices or modules.

Damping Refers to the rate at which reverberant energy is absorbed by the various surfaces in an acoustic environment.

DAT Digital Audio Tape. The most commonly used DAT machines are more correctly known as R-DATs because they use a rotating head similar to that in a video recorder. Digital

recorders using fixed or stationary heads known as S-DAT machines.

(such as DCC)

are

Data Compression System for reducing the amount of data stored by a digital system. Most audio data compression systems are known as lossy systems, as some of the original signal is discarded in accordance with psychoacoustic principles designed to ensure

that only components which cannot be heard are lost.

dB Decibel. Unit used to express the relative electrical voltages, powers or sounds.

levels

of two

DDL Digital delay line.

Decay Progressive reduction in amplitude of a sound or electrical signal over time.

Defragmentation Process of rearranging the files on a hard disk so that all of the files are as contiguous as possible, and that the remaining free space is also contiguous.

DI Abbreviation of direct injection, in which a signal is plugged directly into an audio chain without the aid of a microphone. A DI box is a device for matching the signal-level impedance of a source to a tape machine or mixer input.

Disk Originally

an

abbreviation

of diskette,

but

now

used

to

describe computer Floppy Disk.)

floppy, hard

and removable

disks.

(See

Dither System of adding low-level noise to a digitised audio signal in a way which extends low-level resolution at the expense of a slight deterioration in noise performance.

DMA Direct Memory Access. Part of a computer operating system that allows peripheral devices to communicate directly with the memory without going via the CPU (Central Processing Unit).

Driver Piece of software

main

program

that handles

and

a

communications

hardware

peripheral,

between

such

the

as

a

soundcard, printer or scanner.

Dry Signal to which no effects have been added. A sound which has been treated with an effect is said to be wet.

DSP Abbreviation of digital signal processor. Microchip designed to processes digital signals.

Dubbing Adding further material to an existing recording. Also known as overdubbing.

Dynamic Microphone Type of mic that works on the electric generator principle, where a diaphragm moves a coil of wire within a magnetic

field.

Dynamic Range Range in decibels between the highest signal that can be handled by a piece of equipment and the level at which small signals disappear into the noise floor.

Dynamics Method of describing relative levels within a piece of music.

Early Reflections First sound following a

reflections from walls, floors and ceilings sound created in an acoustically reflective

environment.

Effects Return Additional mixer input designed to accommodate the output from an effects unit.

Effects Unit Device for treating an audio signal in order to change it in some way. Effects often involve the use of delay circuits, and include such treatments as reverb and echo.

Enhancer Device which brightens audio material using techniques like dynamic equalisation, phase shifting and harmonic generation.

Envelope How the level of a sound or signal varies over time.

Equaliser Device which cuts or boosts parts of the audio spectrum.

Exciter Enhancer that synthesises new high-frequency harmonics.

Expander Device designed to decrease the level of low-level signals and increase the level of high-level signals, thus increasing the dynamic range of the signal.

Fader Sliding control used in mixers and other processors.

Filter Electronic circuit designed specific range of frequencies.

to

emphasise

or

attenuate

a

Flanging Modulated delay using feedback to create a sweeping sound.

Floppy Disk Computer disk that uses a flexible magnetic medium encased in a protective plastic sleeve.

Foldback System for feeding one or more mixes to performers for use while recording and overdubbing. Also known as a cue mix.

Fragmentation Process where the storing and erasing of files splits up space on a disk drive into sections. (See Defragmentation. )

Frequency Indication of how many cycles of a repetitive waveform occur in one second. A waveform which has a repetition cycle of once per second has a frequency of 1Hz.

Frequency Response Measurement of the frequency range that can be handled by a specific piece of electrical equipment or loudspeaker.

Gain Amount by which a

circuit amplifies a signal.

Gate Electrical signal that is depressed on an electronic envelope generators and synchronised to key action. designed to mute low-level performance during pauses

generated whenever a key is keyboard. This is used to trigger other events that need to be Also refers to an electronic device signals, thus improving the noise in the wanted material.

Graphic Equaliser Equaliser on which narrow segments of the audio spectrum are controlled by individual cut/boost faders. So called because the fader positions provide a graphic representation of the EQ curve.

Harmonic High-frequency component of a complex waveform.

Hz Shorthand for Hertz, the unit of frequency.

Impedance Can

be

visualised

as the

AC

resistance

of a circuit

which

contains both resistive and reactive components.

Insert Point Connector that allows an external processor to be patched into a signal path so that the signal then flows through the

processor.

I/O The part of a system that handles inputs and outputs, usually in the digital domain.

Jack Common audio connector, either mono or stereo.

Limiter Device that controls the gain of a signal so as to prevent it from ever exceeding a preset level. A limiter is essentially a fast-acting compressor with an infinite compression ratio.

Linear Term used to describe a device on which direct multiple of the input.

the output is a

Line Level Mixers and signal processors tend to work at a standard signal level known as line level. In practice there are several different standard line levels, but all are in the order of a few

volts.

Low-Frequency Oscillator Oscillator used as a modulation source, usually below 20Hz. The most common LFO waveshape is the sine wave, though there is often a choice sawtooth waveforms.

of

sine,

square,

triangular

and

MDM Modular used

in

Digital Multitrack. multiples

to

A digital recorder that can be

provide

a _ greater

synchronised tracks than a single machine.

number

of

Mic Level Low-level signal generated by a microphone. This signal must be amplified many times to increase it to line level.

MIDI Musical Instrument Digital Interface.

MIDI Bank Change Type of controller message used to select alternate banks of MIDI programs where access to more than 128 programs

is

required.

MIDI Controller Term used to describe the physical interface by which the musician plays the MIDI synthesiser or other sound generator.

(Standard) MIDI File Standard file format for storing song data recorded on a MIDI sequencer in such as way as to allow it to be read by other makes or models of MIDI sequencer.

MIDI In The socket used to receive information from a master controller or from the MIDI Thru socket of a slave unit.

MIDI Program Change MIDI message used to change sound patches on a remote module or the effects patch on a MIDI effects unit.

MIDI Thru Socket on a slave unit used to feed the MIDI In socket of the next unit in line.

MIDI Thru Box Device which splits the MIDI Out signal of instrument or sequencer to avoid daisy chaining.

a

master

Monitor Reference

loudspeaker

used

for mixing.

Also

refers

to a

computer’s VDU.

Monitoring Action of listening to a mix or a specific audio signal.

Monophonic One note at a time.

Multitimbral Module MIDI

sound

source

sounds at the channels.

capable

same

time

of producing

and

controlled

several on

different

different

MIDI

Multitrack Device capable of recording several ‘parallel’ parts or tracks which may then be mixed or re-recorded independently.

Noise Reduction System for reducing analogue tape noise or for reducing the level of hiss present in a recording.

Normalise A socket is said to be normalised when

it is wired such that

the original signal path is maintained, unless a plug is inserted into the socket. The most common examples of normalised

console.

connectors

are

the

insert

points

on

a mixing

Octave The shift in range of a pitch when its frequency is doubled.

Oscillator Circuit designed to generate a periodic electrical waveform.

Overdub To add another part to a multitrack recording or to replace one of the existing parts. (See Dubbing.)

Parallel Method of connecting two or more circuits together so that their inputs and outputs are all connected together.

Parametric EQ Equaliser

with

separate

controls

for frequency,

bandwidth

and cut/boost.

Patch Alternative term for program. Referring to a_ single programmed sound within a synthesiser that can be called up using program-change commands. MIDI effects units and samplers also have patches.

Patch Bay System of panel-mounted connectors used to bring inputs and outputs to a central point from where they can be routed using plug-in patch cords.

PFL Abbreviation of pre-fade listen, a system used within a mixing console to allow the operator to listen in on a selected signal, regardless of the position of the fader controlling that signal.

Phaser Effect which combines a signal with a phase-shifted version of itself to produce creative filtering effects. Most phasers are controlled by means of an LFO.

Phono Plug Hi-fi connector developed by RCA and used extensively on semi-pro, unbalanced recording equipment.

Pitch Bend Control message designed to response to the movement of data can be recorded and controller data, although it message group.

produce a change in pitch in a pitch-bend wheel. Pitch-bend edited, just like other MIDI isn’t part of the controller

Pitch Shifter Device for changing the pitch of an audio changing its duration.

signal without

Polyphony An instrument’s ability to play two or more notes simultaneously. An instrument which can play only one note at a time is described as monophonic.

Poly Mode The most common MIDI mode, which allows any instrument to respond to multiple simultaneous notes transmitted on a single MIDI channel.

Post-Fade Aux signal taken from after the channel fader so that the aux send level follows any channel fader changes. Normally used for feeding effects devices.

PQ Coding Process for adding pause, cue and other subcode data to a digital master tape in preparation for CD manufacture.

Pre-Fade Aux signal taken from before the channel fader so that the channel fader has no effect on the aux send level. Normally used for creating foldback or cue mixes.

Processor Device designed to treat an audio signal by changing its dynamics or frequency content. Examples of processors include compressors, gates and equalisers.

Punching In Action of placing an already recorded track into record at the correct time during playback so that the existing material may be extended or replaced.

Punching Out Action of switching a tape machine (or some other kind of recording device) out of Record after executing a punch-in.

Q

Measurement of the resonant properties of a filter. The higher the Q, the more resonant the filter and the narrower the range of frequencies that are allowed to pass.

Quantising Means of moving notes recorded in a MIDI sequencer so that they line up with user-defined subdivisions of a musical bar in order to correct timing errors.

RAM

Abbreviation of random access memory, a type of memory used by computers for the temporary storage of programs and data, and all data is lost when the power is turned off. For that reason, work needs to be saved to disk if it is not to

be lost.

Release Measure of the time it takes for a level or gain to return to normal. Often used to describe the rate at which a synthesised sound reduces in level after a key has been released.

Resolution Accuracy with which an analogue signal is represented by a digitising system. The more bits are used, the more accurately the amplitude of each sample can be measured, but there are other elements of converter design that also affect accuracy. High conversion accuracy is known as high resolution.

Resonance Same as Q.

Reverb Acoustic ambience confined space.

created

by

multiple

reflections

in

a

ROM Abbreviation of read-only memory, a permanent and nonvolatile type of memory containing data that can’t be changed. Operating systems are often stored on ROM as the memory remains intact when the power is switched off.

Sampling

Process

carried

out

by

an

A/D

converter

where

the

instantaneous amplitude of a signal is measured many times per second

(44.1kHz

in the case

of CD).

Also

refers

to the

digitised sound used as a musical sound source in a sampler or additive synthesiser.

Sample Rate Number of times which an A/D incoming waveform each second.

converter

samples

the

SCSI (Pronounced

‘skuzzi’.) Abbreviation

of small computer system

interface, an interfacing system for using hard scanners, CD-ROM drives and similar peripherals computer.

drives, with a

Sequencer Device for recording and replaying MIDI data, usually in a multitrack format, allowing complex compositions to be built up a part at a time.

Signal-To-Noise Ratio Ratio of maximum signal level to the residual noise.

Sub-Bass Frequencies below the range of typical monitor speakers. Some define sub-bass as frequencies that can be felt rather than heard.

Subcode Hidden

such

data

within

information

the

CD

and

as the absolute

DAT

format

time

location,

tracks, total running time and so on.

that

includes

number

of

Tempo The speed of a piece of music.

Thru MIDI connector which passes on the signal received at the MIDI In socket.

Timbre Tonal ‘colour’ of a sound.

Track This term dates back to multitrack tape, on which the tracks are physical stripes of recorded material located side by side along the length of the tape.

Transpose To shift a musical signal by a fixed number of semitones.

Unbalanced Two-wire electrical signal connection where the inner (hot or positive) conductor is usually surrounded by the cold (negative) conductor, forming a screen against interference.

Velocity The rate at which a key is depressed, used to control loudness (to simulate

the response

of instruments

such

as pianos)

or

other parameters on later synths.

Vibrato Pitch modulation using an LFO to modulate a VCO.

Voice Capacity of a synthesiser to play a single musical note. An instrument capable of playing 16 simultaneous notes is said

to be a 16-voice instrument.

XLR Type of connector commonly used to carry balanced audio signals, including the feeds from microphones.

Y-Lead Lead split so that one source can feed two destinations. Yleads may also be used in console insert points, when a stereo jack plug at one end of the lead is split into two monos at the other.