Trials Fusion – “Welcome To The Future”

“Welcome To The Future”, alias “The Main Theme from Trials Fusion”, music & lyrics by Petri Alanko:

Welcome to the future
Man, machine – the fusion
Welcome to the future
Lightyears ahead of evolution

Slay your darkest fears
down the narrow tracks you steer

Welcome to the future

Originally, the song was meant to serve as a backing track for a 3rd party “first look” video of Trials Fusion and Frontier, but it was shelved for being probably too overly pop – or just a wrong song for that kind of usage. Sometimes a song just won’t fit into a picture, or the makers of the video have been listening to a certain song or genre throughout the planning stage, thus planting an earworm, preventing anything from replacing the original. But that doesn’t really end there, instead an evolution (pun intended) set in the way.

When the Trials Fusion soundtrack slowly evolved, took its form and grew some muscles around the shaky bones of the first demos, the people at Redlynx noticed they’d soon need a theme, too – and by then I remembered the shelved track. It had a verse too, but it somehow didn’t fit into the game, so it was left out from the game soundtrack version. Also, the original “alt pop” arrangement was way too lame when compared to the full blast sheer hell provided by the rest of the soundtrack, so something had to be done. Incredibly, it wasn’t the tempo, instead, it was the original version’s rhythm arrangement that sucked most. Having sorted that out with an uneven rhythm basis, the rest of the song just followed naturally. The chorus itself didn’t bring in enough variety, so a soprano was brought to help – this time as a sample, it’s no human voice. To contrast the beauty of epic pads and the soprano even more, three separate basslines were constructed: one providing the body, the other two all necessary flavors of distortion and rectifying.

To tie the far ends of the track together, I used quite a few separate arpeggio patches from Prophet-12 and a John Bowen Solaris – which provides two of the largest pads in the song as well. The other two come from my trusty old Nord Modular G2X, and every patch fitted into the other patch’s holes, although they got a bit help from DMG Audio’s EQuilibrium plugin. A nice one, that. Some of the Solaris’s patches aren’t the most processor friendly, so some time correction had to be made. If outside modulators are used, the machine lags so badly it’s impossible to imagine it’s 2014 – but, shooting out the sounds it does, it’s all forgiven.

Some background vocals were made with G2X, too, using it as a vocoder. I’ve got a nice Sennheiser-ish vocoder patch, which I’ve used many, many times, and again it provided enough warmth and rawness to blend in with the rest of the wailing. Mind you, I’m NOT a vocalist, I just may barely survive singing if done at my pace. Nothing to be heard on stage, that is. I’ll leave that to pros.

The whole soundtrack was made a few things in mind: it should have a few components from the past, and quite a few from today’s pop culture – as well as foretelling what might come in the future. What are the crucial components of, say, drum’n’bass, trance, alternative rock (well, guitar, I suppose?)… and even dubstep? Electro? EDM? What could be the most vital components, the ones that might survive for decades, maybe centuries? We already re-use the Strawberry Fields Mellotron flutes every now and then, Muse had some twang/surf guitar in their huge “Knights of Cydonia” hit, the TR-808 clap hasn’t gone anywhere – and TB-303 just got its digital rebirth in reality as Roland’s TB-3.

Some of the harmonic movements and phrases will stay forever – think of the I-VI-IV-V chord progression, for instance, or basic 12-bar blues structure, or, to put it in chord names, am-F-C-G/B can be found in about two dozen hits within past few years (Alesso’s “City of Dreams”, Armin van Buuren’s “This is what it feels like”, Bob Marley’s “No Woman No Cry”, Katy Perry’s “Unconditionally”, Kelly Clarkson’s “Stronger”… even hipsters were guilty of going there: MGMT’s “Kids” used it too).

So, building on the tradition, extrapolating, converging and simply mashing up stuff in a no-school way was the right route. Usually genre polices are the ones preventing the genre itself from developing, by the way, so a conscious decision was made: at least two different styles must be mashed up in the songwriting, two more in the production. There are a few good examples of that in Trials Fusion’s soundtrack, the few really lucky ones, and the unhappy ones weren’t too bad either, but they had to be beefed up with some extra energy provided by modular synths.

“Welcome to The Future” was one of the happiest accidents on the soundtrack: the female soprano vocals combined with my raw yelling and wailing in the chorus (plus the croning part in the verse NOT heard in the game version), the tranced out synth buildup, the signature digital screeches in the intro, deep unwobbly bass (provided by NI’s Massive and a Prophet 12), and an overly compressed Oberheim DMX with some real trash can loops – oh, and a heavy dose of baritone guitars giving the low edge and rawness. Actually, if looked (listened?) through rose-tinted glasses (headphones?), the rhythm pattern is closely related to drum’n’bass. But maybe people in the future have a bit more relaxed pace of life?

I’ll return to Trials Fusion’s soundtrack in the near future, when time is right, so another post will appear here. Meanwhile: Welcome to the future – the game looks really, really, really nice. And sounds good, too.

Imaginaerum OST – behind the scenes

Now that my official Imaginaerum OST gig is close to its final stages (the 7.1 stems are already in Canada as I write this), I’m taking some time to open up a few techniques used in the sound design and pre-production stage, concerning both the creative method and the application/plug-in side as well. I wrote a rather sketchy blog article focusing on the project as a whole for Nightwish’s Imaginaerum The Movie facebook page ( Me being such a geek – well, a 30% geek, because the other 70% wants to rock with my cock and other pets out – wanted to elaborate the technical side a bit. I’d say the experience level for reading this is somewhere between beginner and intermediate. There’s nothing new or groundbreaking in what I’ve done, that must be said right here, in the beginning.

Plunging into the world of a finished track by means of some Protools project file and everything it consists of, is a really long day’s work. Usually it took two to three days to just prepare the raw material for further processing, do the rough edits, truncate begins and ends, etc. Luckily Nightwish’s mixing engineer, Mikko Karmila, is a hard-boiled pro, who names everything carefully and keeps everything really tidy. The project files were perfect. There were even short notes about microphones used, made by himself or any assistant engineer. It was a peek into a magical world, as I’ve admired Karmila’s productions as long as I can remember, and he’s also one of the very few persons I truly appreciate.

At the listening/selecting stage I virtually had no idea whatsoever what the tracks would turn into. I just listened track after track after track, starting from Jukka’s drums, his overhead stereo file, room mics… everything, and proceeded further track by track. If there was a nice note or a rhythmic thing that caught my ear, I’d cut it off and immediately check the neighbour channels if they happened to share something with what I had just cut out. Sometimes there was suitable audio material, sometimes not. I tried to find patterns, both rhythmical and harmonic. For instance, Marco’s bass playing usually accompanied Emppu’s guitar shredding, so it was natural to chop out similar one bar, two bar, four bar performances from the neighbours as well, whatever there was.

With Jukka’s rhythmical stuff, I bounced some stems from his separate kick-snare-hat-toms-cymbals(overhead)-room-whatnot tracks and transferred the stems into Ableton Live. To those non-musicians, Ableton Live is a clever piece of software that allows separate handling of time and pitch, unrelated to each other. Usually, if you’re lowering the playback speed, the pitch of the signal will lower, too. Just like on the cassette or a vinyl. (If you’ve been around long enough to use such machinery.) That speed/pitch treatment could be done on pretty much every other tool or DAW as well, Protools among them, but Ableton Live’s got many nice built-in tools and plug-ins for mangling the audio stuff beyond recognition. My favorites for Live drum/percussion treatment are Corpus and Resonators. I also did some MaxForLive programming for weird higher-end audio treats. Stuff that mangles stuff, you know.

A good mixture of Resonators and Max stuff can be heard in the Last Ride Of The Day, which will be named differently for the movie and the OST. A rollercoaster usually makes a ticking noise when speeding through the track – so immediately I thought about taking Jukka’s overhead, high-pitched, crystal clear cymbal/hihat/leaking drums and chop them into 1/16th notes, carefully fading out the end of each hit. Added a resonator to enhance the metallic tone – then put it through Corpus to add some “noise components”. Then added some multiband compression to tame out wild lower middle frequencies. That was probably one of the easiest things. Usually the process got out of hands when I just kept going further and further for hours, just playing around. It didn’t take too long to prepare “a path” inside your brain – you heard a raw track and immediately figured out what it could be turned into. After the first week I was literally swimming in files. There were thousands of them, everywhere. Folder after folder filled up. If I prepared a pitched loop, I’d prepare every semitone as well, so if the pitch was supposed to spread over two octaves, I made a separate sample for each key. Transposing a sample sounds worse than playing a specifically prepared sample for that particular key, in my opinion. To make things even more complicated, the pitching was seldom enough. Usually a blend of Ircam Trax in conjunction with some other esoteric plugins was implemented. It was an amusement park, truly.

I really loved putting toms and kickdrums (and taiko as well) through Corpus, turning them into huge bass synths with a percussive envelope. Arabesque is one of the tracks that utilised that technique. A booming, pulsating groove underneath the ethnic percussion… sweet. Delicious. If Corpus only were a reliable when it comes to tuning. Let’s just say a lot depends on its input. Tough transients make it bow and bend, I’ve noticed. When these sub-bass pitched components were mixed in with the original instrument, together they created something wonderful. I knew I was to walk the right track.

Emppu & Marco faced similar treatment, although the tools were somewhat different. I sometimes removed all noisy components (as in tonal versus noise) from the signal, leaving only a mass of sine waves – or rather, a shred part without the harsh edge… would it be called “fredding” then? The outcome of this process has thousands of “glassy” artifacts, which sometimes are delicious, sometimes not. I went a bit far with some guitar stuff, as I pitch-perfected each string separately, and not only that, sometimes I removed all pitch fluctuation from the low A and E strings to “keep the package together”. Of course, after reamping, it was no longer a guitar as such, instead, it became almost an ethnic instrument due to the strange flavor it got in the tuning process. Luckily the stringed instruments provided me with enough material for some of those really low brass-sounding stabs that echo in, for example, the teaser. Of course, those ROAAARR stabs are always doubled (or rather, quadrupled) with something else, as it’s all about texture. A bit like looking at a really good picture: if you see a beautiful portrait, you probably see the skin first, then you pay attention to skin glow, irregularities, maybe hair too, the makeup, clothes, lighting, posing, environment, dust in the air, the moustache on a woman… it’s not “a portrait” of a hairless naked transparent albino cave olm floating in space, it’s an interesting person in a realistic setting realised by professionals. The final work of art consists of thousands of details; layers.

Usually using just “a sample” creates that olm feeling. Layering two sounds makes just… a pair of olms. I’ll return to this a bit later.

Sometimes the two shredding masters of Nightwish created a groove that had a magnificent “suck” to it, their playing was literally in a pocket. Unfortunately – forgive me for saying this right after praising their playing – the riffs they played quite often employed notes not necessarily feasible for transposing, all the semitone/tritone jumps. They are incredibly effective for pounding the stuff through in one key, but I couldn’t use them as such, I needed to cheat a bit. During the listening/preproduction stage, I usually start forming a track inside my head, not necessarily knowing what I’m doing – it’s an almost subconsious process, where I grab a chorus chord progression and try to fit in another part of the song, which often was in a different key. It’s a process of “I could use this. And this. Maybe that too, and those”, picking up what I loved in a certain song – or songs – then glueing them all together by means of technology. And often all of the aforementioned tracks/stems were in a totally different key. Including the orchestra. And the ambience channels, too.

Also, for creating the riffing guitar/bass parts, I had to create straight riffing/shredding patterns out of the original parts, which included a lot of jumps. Which had to be taken away, still retaining the original groove. I just had to cut off the “wrong” or “unneeded” notes and replace them with surrounding chugs/notes from similar subpositions elsewhere. By just replacing a second 1/8th note of a second beat with something from a different beat destroyed the groove. Definitely not a good idea. So – I just had to listen to the tracks really carefully and find exactly the right beat, right note with which I could cover up the hole left in the original groove.

By now you start thinking “why the hell didn’t he just call in the guys and have them play it for him” – but it would’ve been against the concept. And timetables. When I started pre-production phase, they had different engagements to the album, tour preparing, vacations… there were many different things. And – my luxury – I had plenty of time to mangle the stuff beyond recognition. Why wouldn’t I do that if I’m allowed to? Projects such as Imaginaerum OST are a rarity. When I come across one, I use it for studying as well, as I’m constantly eager to learn new things.

After the holes were filled, I needed to create realistic transposing for the gtr/bass patterns. Also, as I really wasn’t too sure in which tempo/tempi the certain tracks were setting into, I needed to have the shred patterns disconnected from their original BPM. Oh crap. Enter Native Instruments’ Kontakt 5 – a brilliant, brilliant sampler plug-in, a piece of software that works as a stand-alone application or, which is how I use it, as a plug-in inside my choice of DAW (digital audio workstation), Logic. Their previous version, NI Kontakt 4 had some issues with transpositional artifacts, especially when using the more specialized versions of the sample playback engines, Tone Machine as well as Time Machines 1 and 2. Luckily, their version 5 fixed nearly all of the bugs, and their newest Time Machine Pro sounded realistic, even with signals containing a lot of distortion. After a few short tests I was convinced with the results, as the transposed sections blended in with the untransposed rather nicely. I did a few artificial transposed versions to extend the tuning ranges for more… well, dramatic sections. Drop C tuning was still a bit high at times… I was afraid I’d lose the coherence of the low end in Marco’s and Emppu’s playing, but with some help from my Symbolic Sound Kyma system, I was able to create a bunch of murderous, angry tones which kept their tuning really, really well. Kyma is a hardcore DSP scripting/processing/creating environment, virtually a tame black hole. You can easily lost days whilst noodling with it, just testing out a few algorithms. All the most memorable sounds in the 90’s and 00’s and 10’s are – almost without exception – treated with Kyma. With a reason. That… thing is like a superhero amongst humans. The quality of its output is always above all expectations (greetings to Carla Scaletti and Kurt Hebel), even though the user were an idiot. I know what I’m talking about.

After the additional transposes were finished, it was time to put together a virtual Emppu/Marco entity. EmMa? Marppu? Emco? Just dragging the right samples onto the right notes in Kontakt 5 and spreading the key groups gave me a reasonable assurance; yes, this works. I added a round-robin feature (which randomly selects a different, yet similar sample from a specified stack of samples, thus preventing a stupid machine-gun-like effect usually associated with badly prepared sample sets) and some volume control, filtering, even a slightly randomized multiple eq settings (only +-0.5 dB with narrow Q values)… every trick I knew – and it sounded even better. The finished EmMa (sorry, guys) allows a rather large tempo adjustment, +-30 BPM away from the original 157.0139, so I could use it in other OST cues as well. Brilliaaaaaant!

Making a virtual bassist from Peppepappa on Vimeo.

Note: similar approach was taken with Emppu’s guitars. Nobody was left aside. Kontakt 5 can be seen in the latter section of the video.

But, without the true passion of the original performance, it would be nothing. Again I must emphasize the fact that the tracks oozed with artistry and passion, yet they were executed with immaculate professionalism, no glitches whatsoever. It was a set of every sound designer’s wet dream; not just a bunch of tracks for raw material, but a bunch of tracks with a Meaning. They had had a lot of fun during the recording, I’m sure. You can hear it.

Whenever retuning was ever needed (a rarity, unless I was after some effects), my choice of weapons were Melodyne (which produced quirky end results at times, probably due to complex input signals), AutoTune (for extreme quirky effects), Reaktor (grain cloud stuff and eerie stuff) and Zynaptiq’s PitchMap, which thankfully came out just in time before the final production phase. The last one, PitchMap (PM) got used a lot, in some song structures there were about a dozen or so instances. PM allowed me to change pitches polyphonically in realtime, without preprocessing. I wish every DAW application had PM as a built-in tool, it’s that versatile.

Another trick I used often was Melodyne’s polyphonic correction tool (when it worked). I sometimes bounced an orchestral riser (a crescendo) off the Protools projects and imported it into Melodyne and had it analyzed with the highest possible setting, then meticulously retuning the lower notes onto absolute notes separated by an octave (or two). Even though the original riser was dissonant, the retuning brought it back to normality, but I was able to control all the harmonics with the retuning process, sometimes removing the pitch fluctuations completely from certain frequencies, sometimes exaggerating it by 100% – as much as I could. The results are heard in the first reel of the movie, in the second open-air scene with the main characters. If you can hear a sound resembling a bomber plane screeching through the clouds at about 13 minutes and 44 seconds from the start, it’s an orchestra. Brass section. Note: as these are “work-in-progress” files, you’ll hear a lot of artifacts. The final files are in the Imaginaerum OST.

I often doubled the EmMa stuff with staccato or spiccato strings, or, celli and contrabass sections. I tried to use the original orchestral tracks as much as I could, but there were times when section and instrument leakage was so heavy I had to recreate it with samples. The leakage is unavoidable; there are dozens of players in the same room, all blasting off at the max volume, so they’re going to leak into each other’s microphones. I think they’d leak even though contact microphones were used (which grab the signal through a resonating solid material, not through air), the sheer volume of everybody playing forte fortissimo makes everything resonate; brass, wood, strings, cymbals. Everything. The leaking makes a part of the sound, it unifies the sound field, melts every single instrument into one big instrument: an orchestra. Without leaking it would sound like a badly done synth orchestration, sterile and embalmed. Dead. Leaking & dirt = good. Just as in real life: good sex is always a bit dirty. I think I’ve said that before…

After some heavy testing, I chose LASS (Audiobro’s Los Angeles Scoring Strings sample library) to accompany the shredding. Sometimes I emphasized the sound with some other sample sets, even with some staccato low end brasses as well, but about 70% of occasions, I managed with LASS alone – although with several layered sections, doing slightly different things.

Yep. Layers. There are a lot of them. If there’s a pad, it’s not just one pad or blast or choir or whatever, it’s a choice of three, four, don’t-know-how-many sounds each playing a bit different part. The more further away the notes of a certain part are, the quieter it is. Often thinner, too. Sometimes I layered a “resin” sample with a granular choir, to make it sound a bit string section-like. There was a lot of controller info, sometimes affecting a parametric EQ, sometimes the amount of granularity (i.e. more granularity creates a decomposed sound made of bits and pieces of the original signal)… and always the layers were tightly tied to certain frequency areas to keep the whole patch easily playable. Actually, now that I’m looking at the instrument sets I’ve done, I’m quite convinced that almost everything could be played live with some careful planning, but it would require a lot of work and a large sampler or a workstation. Maybe two or three.

Here’s a real-time demo á la Imaginaerum OST (just a noodling, improvisation):

As I said earlier, I’m more player than a geek. I’m very fond of instruments that are simple to create with, which is my “secret weapon”. One can have the most incredible sample library at hand, but without at least a decent usability it’s worthless. I seldom take sample libraries as is, usually I’ll spend at least a month or so creating my own set of instruments which I can handle easily, without extra thinking. In a creative are such as mine, it is imperative to know your tools and how to get everything out of them. You must concentrate on the creative side and raise above the technical issues. You must be able to think quickly, without a hinder. You must be a mad geek to tame all the power, but you also must be able to express your passion through playing. The tech is there to be conquered, it is a good worker but a lousy master.

I must admit I felt a bit embarrassed to treat Tuomas’s keyboard tracks in any way. He’s an excellent player and just being able to dissect his compositions felt a bit odd at first. If possible, I kept his keyboards in every cue as long as possible, but since the original performances were audio only (no editable midi data available, thus preventing me from using his playing with other sounds, other than recorded ones), I faced a dead end every now and then. For instance, if a cue needed a softer piano or fewer notes, it didn’t exist. It had to be recreated. To an extent, I could use Melodyne Editor. It took most of the piano parts without pain, but just as soon as there was some harmonic or phase movement, it ran into problems. Luckily the aforementioned PitchMap arrived in time. However, if I needed to create a midi file from Tuomas’s playing, it had to be done with Melodyne, which can output the notes contained in an audio file into a certain type of file that can be used in every DAW application. The moral? Screwdriver isn’t enough. A carpenter needs a saw and a hammer too. And quite a few other tools as well.

Sometimes even that, the whole toolbox, didn’t do the trick, and I had to play a lot of stuff in myself. I aimed for maintaining the original voicing and the choice of intervals and chord inversions whenever possible, but at times I had to create a section from scratch, or the movie needed something that couldn’t be found on the Protools projects. Needless to say I was overnervous when sending a such track to Tuomas and Stobe (Harju, the director of Imaginaerum, who also wrote most of the screenplay), because if there had been anything not worth a shit, they’d surely kicked my ass immediately, although politely. Think about Alan Rickman saying “how grand it must be to have the luxury of not fulfilling the expectations. Now go back to your pathetic chamber and concentrate like your life depended on it. Which it, of course, does.” To be honest, none of that ever happened, but I sure felt like a wand-waving apprentice at times. Turning gazelles into frogs, or, with some luck, into unicorns.

Of course, the processing didn’t end there. Sometimes I had to get rid of trombones leaking into everything else, sometimes – although it was a rarity – flutes were playing on top of a perfectly usable high violin/viola motive. Eventually I started questioning my sanity, and especially the methods I was using? Did I really have to do this and that. I spent literally dozens of hours with Sonicworx•Isolate, which allows one to “draw off” unwanted signals. By December, it was really painful. I found errors in my versions, things that didn’t fit in, or was just lame, repeating itself, or sub-quality in my opinion. Or did I just imagine everything? The project started to haunt me, and a short while I was really angry to let people hear something I wasn’t sure of. During the first two weeks in December I re-did a lot of stuff, partially due to the first finished cut of the movie, as there were a lot of off-beat picture changes which I had to take into account, make tempo changes, re-arrange etc. Slowly, the versions were reformed and it was as if a new backbone had been installed. Finally, the clicking into place started.

I’ve used the term loosely here, “clicking”. But it describes perfectly the feel I was after. Puzzle pieces should just click, they shouldn’t be put into form with force. If force was used, the pieces would break and render unusable, creating just a mess. I felt the score needed a few lighter versions to counterweigh the angry and sinister side. Mermaids and CrowOwlDove happened that way – actually, even though COD is a shorter version than its original, it is one of my favorites, as the chorus/verse parts were mixed, and the song structure really builds into something else. It was a song virtually conceived by mixing up the pieces and rebuilding it blind. Or rather, deaf. I just looked at the previews, the sausages – or the continuous wire waveform – and cut the pieces without having my loudspeakers on, just trusting my instinct. The end result was surprising. Of course I had to hone out a few rough ends, but it did sound rather good. I fixed a few fighting notes from the lead vocal and that was it.

If I should explain why I chose the parts I chose, I think my most overused answer would be “I felt that one would be a nice one”. I made those choices very, very quickly during the first few listening rounds, just ran blind, trusting my gut feeling. There were bass hits, single drums, percussion rolls, orchestra blasts/crescendi/whatnot – and sometimes I chose just noise, noise from the orchestra’s ambient channels. Sometimes a chose a leak, say, trombone leaking to doublebass channel, or vice versa. I wasn’t after the most beautiful single note or phrase, I was constantly looking for dirt and interesting, fiddly parts that would benefit a lot from corrections and/or editing. A lot of ambience stretching, denoising and retuning was done to create the surreal pad instruments often decorating the simpler scenes. Of course, there are a lot of commercial libraries and software instruments involved, as well as my trusty hardware gear – especially my modular synth, which processed a lot of the tracks and even got to be used as an ambient piano in the beginning scene! There are a lot of lower midrange growls from my trusty old Oberheim Xpander. Some guitar parts were doubled with a Dave Smith Mopho Keyboard, a tiny monophonic analog synth, as its sound structure is just perfect for my use, and I can get it sound like an angry banshee on a charcoal grill.

More about bad samples: Just using a good snippet would be comparable to taking a picture of Mona Lisa’s shoulder and glueing it into your canvas. Then repeating that, until you’ve got a Mona Lisa Polaroid Puzzle. It could be interesting, but the outcome would still be merely a bad copy of an original masterpiece. My philosophy was to sketch Mona Lisa’s shoulder onto white paper with a pencil, then take my polaroid and mangle the polaroid beyond recognition, creating a “multi-technique collage” that would possibly have a life of its own, disconnected from its origins, yet nodding towards them with respect. I just wish I could’ve used those funny and dirty sounding eerie pads even more, but the movie had almost an action vibe to it, especially closer to the end.

And that vibe needed a healthy dose of percussive instruments. Some of it is from Tonehammer, which later on became 8dio and Soundiron, some of it is my own library I’ve recorded throughout the years, as I began doing this when I was in high school, some of the samples I use even today were originally sampled into an Ensoniq Mirage, back in 1986. Oh, the agony of not having enough memory! Things are quite different today. Sometimes I needed to employ three computers, one took care of the percussion, another took the strings and main machine did the rest. They were connected by means of Vienna Ensemble Pro 5, which allows one to connect multiple computers with an ethernet cable, carrying midi and audio signals. During December, I installed multiple SSD drives into my system, removing the “your computer is too slow” message almost completely. I regret I didn’t do that earlier. Not every cue required three computers, though, especially after the SSDs entered the house. About 95% of final mixing/editing time, one Mac Pro was enough. It was only the most crowded string and percussion arrangements that needed a few extra hands.

I personally happen to like lower frequencies a lot. With some careful programming and arranging one can cause the toughest cold turkeys, making each hair stand like they were dropping off suicidally. In the autumn 2011 I began experimenting with piano and harp, as well as brass instruments, which I noticed benefit from pitching down their tuning. I collected a lot of lower midrange and bass instruments from my libraries and Nightwish Protools sessions, and created some hybrid instruments based on the lowered samples. The results? Let’s just say the tripods from The War Of The Worlds remake were their little brothers. Also, it was a respecting nod to the other epic direction: Inception. It was almost obligatory to incorporate keyboards into the sound, both Tuomas and myself being keyboardists. I also happened to have a background of playing pipe organs (I probably had the best teacher in the world – whom I didn’t appreciate at that time, unfortunately: Kalevi Kiviniemi), so I put in some of the lower, richer reed stops that were recorded in Lahti in a church back in the early 90’s. I’d definitely like to re-do that session, by the way. That and a grand piano. And a harp. And EmMa. And… a Novachord? Actually, not a real one (as it’s beyond most people’s budget).

Improvised á la Imaginaerum OST:

(By the way, look for Kalevi Kiviniemi’s version of Gabriel Pierné’s Prélude, Op. 29, No.1 from iTunes and let it play – it grows slowly. By 1’35” it turns into a gothic masterpiece. THAT is how you play organs, note also the grand, lush ambience of a gothic church it was recorded in. The G chord ending that one is something unreal. I’ve been practising a lot lately, so I might match him as early as 2040. Maybe sooner, with some unexpected luck.)

With some songs I regretted the order in which they had to be incorporated into the movie. For instance, one of the “underused” tracks was Song Of Myself, an epic that should’ve had more room, but alas, the script was changed and most of it had to be taken away. The most beautiful part of it is still there, luckily. Without giving out any hints of the plot or the script, I’ll say it underlines one of the most beautiful and wistful moments of the movie, with a firm yet tender voice.

I used a lot of choir and vocal takes from Imaginaerum The Album – but left most of them alone, choirs weren’t processed or turned into dinosaurs wailing at the peak of their climaxes, nothing like that. I felt it was important to preserve yet another element from “classic” Nightwish – although I created a monster monk chanting crowd based on Marco’s incredible voice. Anette’s lead vocals appear here and there, and in a few occasions one lead vocal and its doubles were turned into a choir of 30+ vocalists. Her voice, however, was so clear and pure, that it felt a bit awkward to mutilate it in any way. A bit like employing some bradpitt and putting a monster mask on him for the whole movie – what’s the point?

And no. Brad’s not in this film.

If you’ve really read this far, I’m nothing short of amazed. WIth all these references to technical issues, applications and plugins, text becomes easily stodgy and onerous. I’m sure I’ll write a few more lines in the future – maybe dissecting a few scenes sonically, but it’s a totally different story then.

Cheers, enjoy the ride!

It’s been a while

Time does fly. Seems that there’s been just too much to do. Always a good excuse.

I started working on a movie score a few months back – a movie more or less based on a yet unreleased album by Nightwish (it’s coming out in 2012, I think). The screenplay and the whole plot is just simply brilliant, and Mr. Stobe Harju, one of the manuscript writers too, directs it. You probably knew every word of this already, if you’ve been following either the news, Tuomas Holopainen’s or Stobe Harju’s twitter account or visited the Nightwish Official Website. The whole concept of this score is a bit different from everything else, a bit dogmatic as well, some might say: I’m using Nightwish’s album multitrack data as my raw material – and only a little else. I’d say the percentage is 80/20 for NW, clearly. There’s just so much to build onto, so many building bricks. The only evil thing is the toughest question I’m facing on every track: where the hell to start from? What’s the ground zero for each song? You could turn every song into a duck – or a unicorn, easily. Every piece has enough stuff for that.

I turned out every song had two or three “hooks” that seemed to talk in a clear language. They quite seldom were melodic motives, though, usually they were just some odd noisy bits and pieces that immediately triggered the train of thoughts – and sometimes I just listened to a whole orchestra track thru a double bass channel, in mono. Thru an unmerciful compressor. After making very comprehensive memos of everything, every track, every song, I had a clear picture of what’s where. The initial transfer from ProTools to Logic took a month – of course, I made my notes at the same time. Also, I created a set of Kontakt 4 instruments on the fly, grabbing a straight note there and another from here, the usual sample stuff. I’m not in my comfort zone now, as I’m depending on solely someone else’s work of art, and am willingly, knowingly, trying to steer away from the I’m-replacing-everything-with-my-output easy solution lane (which everyone, including me, would hate eventually). Everything gets treated. Really, really mangled, to find the right patina for the existing colour. I’m not destroying Tuomas’s brilliant work, I’m building something new using existing material. Call it reinterpretation, whatever, but I can guarantee they’re not just mere remixes. And there will be no dance beats. (At least not yet.)

(By the way, as a sidenote, I’d like to encourage the software developers to finally fix the frigging, effing, bugs that prevent the AAF transfer from working. Thank you. It’s not working no matter everybody tries to say. Not at least from ProTools to Logic.)

So, I’m in “under construction” mode right now. It’s a funny state of mind in which I seem to be unable to find time for shaving alone. Also, during this phase I’m also consuming vast amounts of green tea and find sleeping a hard job. I’ve been thru these phases so many times that I’m already adjusting my mind for the misery of giving up, leaving the project, you know, the moving-on-phase. Luckily, it’s still months away and I’m enjoying the energy, but I recognize the pattern.

Also, before I even started about thinking about starting to think about thinking about Imaginarium, I laid my hands on a few game soundtracks and in addition to those, I also had a chance to collaborate with my #1 person to collaborate with, DJ Orkidea, on his upcoming album. I even wrote a series of blog articles for every single track I was involved with, but decided to postpone them until the release gets closer. It’s going to be some really, really hardcore nerdy read, but I feel that opening my bag of tricks is not equal to letting my mind slip away from me. I’m using my tools my way, according to my taste – someone else will use the same exact tools in a totally different way. One shouldn’t fear sharing the knowledge (if such thing exists).

I’ll try and update this site accordingly when the time comes.

“Thunder & lightning, very very BZZZZaargh”

Last week a thunderstorm went over Southern Finland – and without much further thinking I ran outside with a mic boom in my hands, only to remember a short while afterwards it was really a bad idea, as the boom stand was metallic, as was the mic… and we have no lightning conductor in our house – well, effectively I was that for short while – so I decided to continue with only the internal mics of a Zoom H4n. Not an obvious choice, but I didn’t want to soak my fake foam head made for OKM ortophonic microphones, and the OKM mics themselves.

So here you are, a nice August thunder in the countryside: Last week a thunderstorm went over Southern Finland - and without much further thinking I ran outside with a mic boom in my hands, only to remember a short while afterwards it was really a bad idea, as the boom stand was metallic, as was the mic... and we have no lightning conductor in our house - well, effectively I was that for short while - so I decided to continue with only the internal mics of a Zoom H4n. Not an obvious choice, but I didn't want to soak my fake foam head made for OKM ortophonic microphones, and the OKM mics themselves.

So here you are, a nice August thunder in the countryside:
Ukkonen 100810 44100 by alanko

Another review of Alan Wake score

Catherine Spencer (of melted my heart, literally: “A Winter’s Dream sets us up for forthcoming disasters and sets out Alanko’s stall – a sweeping epic backdrop to play against. Although never becoming choral, these recreate similar emotion to Howard Shore’s Lord of the Rings soundtrack.” She manages to mention both Howard Shore and Petri Alanko in a same sentence – which is about as close to heaven as one can imagine, sort of.

Mr. Shore made about 10 million people cry in a nanosecond with the first notes of his Lord Of The Rings masterpiece soundtrack score, so – hell, if that’s not a good comparison, I don’t know what is. 😀

Read the rest of the article here.

Alan Wake music ratings

It really, really makes one humble to notice someone else digs what you’ve done with your heart bleeding. Seems that “Alan Wake” is a springtime black horse in the audio/music field as well. “The volumetric fog, light, amazing sound design and of course the absolutely incredible soundtrack Petri Alanko has composed for that game, really does bring together an incredibly eerie feeling.” “Remedy is really pushing surround use and LFE (bass) response to build tension and create fear as new “phenomena” are discovered, while the score by Petri Alanko is appropriately spooky. Those of you with a 5.1 system and a decent subwoofer will greatly benefit.”

Jeff Talks Games: “Petri Alanko, who composed the score, deserves some serious praise for being able to set the mood and tone of the game. There were times I realized I was sitting on the edge of my couch even though there was nothing going on with regards to action – it was all based on the music.” “(In Googlish:) On the other hand, the game’s soundtrack is provided by Petri Alanko. This multifaceted creative is driven by a variety of genres throughout his career: Alanko compositions created for Alan Wake (emphasizing the main line) are of astonishing quality. Without any doubt, worth checking out the career of this unknown composer, to realize that there are true geniuses in every corner of the planet.” (check the original article in Spanish)

Also, Finnish Pelaaja magazine gave Alan Wake quite a warm welcome – and they even remembered to mention music, which is a rarity among Finnish game magazines. “Tunnelmalle antaa paljon myös Petri Alangon upea musiikki”, they said, which could be translated as “The overall atmosphere benefits from Petri Alanko’s great music”. Try that in Google Translator and prepare for a disaster.

What can one say: Thank you, very, very much.

It’s been a while…

As I mentioned in my last post, it takes balls and guts to maintain a blog. I haven’t forgotten mine, but there’s been too much to think about. Now that everything’s done, I guess I’ll have some time to write a few things here.

But first I suggest you go and check this. Mainly because it’s cool, but also because it’s a clever ad campaign / real life prequel for Alan Wake, the game for which I composed music. There’s also some music of mine in the Bright Falls series as well, most notably in the end of Episode 2. I guess they used it in the other episodes as well, but… time tells.

The tough challenge

Damn. I once said to a good friend of mine it takes big balls to maintain a blog. I’ve always hated someone just filling the space by babbling, but unfortunately I haven’t done even that. Even though there’s something happening with the projects I’ve involved with, there’s not much to say – other than: Yes, everything’s under construction. I’m just one frigging lazy piece of shit. Oh well.

Back the other day I calculated the amount of music needed for a certain Remedy Entertainment game (AW) and practically froze: The sheer scale of the emotions inside the product is just, well, huge, to put it plainly. From absolute horror to despair to intimation – whatever and everything in between – demands a few good night’s sleep before even thinking about booting Logic. The music has to be alive, it has to breathe, it must take the player into the world he’s seeing in his eyes. The music must function as pylons of a bridge created by gameplay, a bridge between the real world and the fake one. That demands quality, quality demands time and concentration, which in turn demands peace of mind – and the ability to emphatise. It may sound damn tragic for a person of my age, but I’ve cried many, many times during the sessions: because of the scale of the environment, of the characters’ emotions, of the air you can see (hell, you can see the temperature as well, not from a thermometer, but just looking at the air and the colour of the light, just like in the real world). I honestly don’t know whether I’ve managed to interprete the feelings involved, but it’s a good start. To feel. To feel something from a mute rendered picture, that is. People at Remedy are very skillful in delivering the finesses, I must say.

Which leads me to a problem that virtually every composer has encountered. The UI finesses of orchestra libraries. Or lack of them, actually. I own some of the Vienna stuff and two licences of EastWest’s Symphonic Orchestra Platinum ProXPwhatthehellitscalled plus some more esoteric stuff and a huge 80+ gigabyte library built during the last… wait a minute… twenty years or so. I still today use some of the Huge Drums library I built with Emax and Akai S-1000 in the early nineties. Even though they were originally recorded in 44.1 kHz and 16 bits, they were recorded to my best knowledge by then and they contain some of the meanest lows and rudest transients ever encountered. They will cut through.

Anyway, back to the lack of UI: before mastering any of the library engines, a serious amount of practising is required, just like becoming a virtuoso instrumentalist. The required time spent with libraries is not that long, of course, but requires similar attitude: you have to accept you’re producing plain shit in the beginning. You shouldn’t practice whilst working, instead, you must spend enough time to be able to deliver things in time when needed. One can do decent stuff after fifteen minutes, yes, stuff that fools people for 10 seconds. But it doesn’t breathe, it’s a mere polaroid of a living thing: an orchestra. Every time I hear a sample orchestra I visualize someone throwing Polaroid pictures in the air, instead of someone running around, lustfully painting on a huge canvas with lively, real, colours, using brave movements of hand and body – which is how I picture a real orchestral entity: a group of skillful musicians all controlled by the brains; a conductor, that is. All the usual UI stuff: legato patches, portamento, switching keygroups, whatnot, it just doesn’t work. You just switch the Polaroids, that’s all. Human brain is a wonderful thing: it recognizes patterns, no matter how dense the mayhem. The first 10 seconds are fine, then your brain kicks in: “Ah, THAT attack”, “I must’ve heard that vibrato before”, “Funny that those strings are bowing in a constant speed even though it’s tempo rubato”. Even though the listener’s not a specialist, his/her brain figures out there is repetition.

I tried using a solo cello patch from EastWest’s Play edition of EWQLSO, and was so deeply frustrated I could hardly bear myself. Of course I kept on practicing, trying to master the software, but it didn’t produce anything even remotely decent (yes, a living cellist would’ve done it in a single take). Turned out their engine had some serious issues (serious for me, not for EW, because they had put the library out without finalizing the patches). I contacted their tech support, which took a few days to investigate the problem, then sent me a corrected solo cello patch for the piece I was working with. That did the trick and I could finish the track in time. However, their patch of a patch saved only the Solo Cello I was using, not other Solo Portamento instruments I was planning to use. And the other solo instruments were detuning themselves pretty badly as well, a bit like yours truly playing a viola (which I cannot do at all), so I had to give up. Portamento engine malfunctioning or something. Now I’m waiting for a total update. Hope it arrives in time.

I haven’t given up yet. I do believe it’s possible to create a decent string orchestra that can be played with a keyboard and doesn’t sound repetitive or synthetic. It’s just a matter of choosing the right controllers to do the right stuff. Vienna is pretty close with their speed-of-playing-switching -thing. Unfortunately, their Player plugin (or application, actually) is somewhat flaky for a person that has used AU plugins for… too long, I suppose. And unfortunately it’s not possible to do crossfading. Or if it is, my Vienna player is definitely malfunctioning as I’ve yet to come across with a suitable solution. Yep, I’ve read the manual pdf. Many times. I even printed it on paper and studied it thoroughly everywhere I could. One thing is sure, though: keyboard velocity as we know it is not for me, at least not in a normal 1-127 scale. I prefer to scale it down to 30-100, sometimes even 45-90, and connect the expression pedal to partially control the velocities in Logic. Which is hell, by the way. 🙂 Logic’s buggy, you know.

So if we multiply the bugginess of the mother app with plugin issues and scripting issues, we’ll confront something the size of the huge robot beast that’s tearing down the Giza pyramid in Transformers 2. Add that to the fact that you’re actually trying to do emotional stuff, to compose, to emphatise and move people, to create original stuff without cloning the same theme all over again… you’ll face quite a few challenges, that’s sure. And all of a sudden your deadline is no longer a tiny thing in a distance.

The best thing: I can build on these feelings. The anticipation, the fear, the nightmares, the frustration, the anger. Everything becomes one, it’s all going to be tiny bits of Alan Wake score. The destroyed mother keyboards (two gone already), trashed mouses (two), angry emails (literally dozens). I could say software bugs are creating part of the score, whether I wanted or not.

Do I have to share my credits as well or did I acquire a licence for them as well when I bought the libraries? 😉

Päivitys / Update

Päivää. On ilmeisesti syytä päivittää hieman blogin nuutunutta ulkoasua ja samoin tein muuttaa kieltä kohti sointuvampia äänteitä, kiitos erään klasariprojektin. Lisää seuraa.

Ja kommentit ovat edelleen poissa käytöstä tammikuulle 2009 asti.

And same in English. Hello there. It is time to update the rather dated look with a new theme (thanks to David Garlitz for that) and switch the language – at least part time – to English, mainly due to a certain pseudo-classical project released by Armada Music. More to come, hopefully sooner than later. Also, I’ll be updating this retroactively and bring in some blog writings done elsewhere. I suppose they’d fit in nicely, them being all tech talk and about modular synths, microphone disasters and digital whatnot. However, due to readability, those posts are going to stay put in their original date positions. As my non-native English roots force me to think as a translator, the language I use is probably a bit harsh and somewhat demented for some of the more distinguished bypassers, but I’m sure the train of thought will still be clear enough. Most of the writings in the Diary category are usually project related data – and definitely genre-free – and mostly concerning electronics and sound design stuff as well. That’s what I do. Sound design.

Oh, one thing. Unfortunately, until January 2009, I’ll have to keep all comments off.

I’m hit, I’m hit!

No, wait. It’s just my MacBook. Or actually, not mine, it belongs to a friend of mine, but I’m sort of responsible for the damage.

It was Midsummer Eve – which is a one huge party all over Finland (and Finland’s the place to be then, no matter whether it’s raining your ears off or scorching heat) and I was gigging. The roadies were taken a few beers too much and forgot to fix the stage clamps, which made the whole stage rig one big trampoline. If the guys in the front jumped up & down, the stage behaved like a flooding tsunami. Waves, literally.

Then the rack it was placed on started to move. I noticed it crawl, but had my hands full, playing as I was. I watched silently MacBook fall down from the rack top, landing on its screen. Mysteriously, both the USB and FireWire cables were still connected and backing tracks running. Phew, I thought, it just scared us.

The song ended and I had some seconds to move the laptop back to its place. I froze. The screen was gone bad. Actually, it was crushed.

It was devastating to see how the gig would end – or would it? There was a tiny fragment of display still intact, about 1.5 cm x 8 cm, and if I used my finger carefully, I’d be able to select the right song file: I had printed a song list on the paper, which I never do. So, if I moved my finger 4 mm down, the mouse pointer moved down one song slot, selecting the 10th song means I’d have to move it 4 cm down. Press Space to start, listen to click. The tempo sounds familiar, maybe… Yes! 🙂

The main reason for the accident was, as it turned out later, not remembering to place the soft cushion-like piece of 3 cm thick polyurethane underneath our live MacBook. A mistake almost costing us our reputation as a live group (Ableton Live group, that is) and with some bad luck, the gig would have been over just after one song.