… of seeing a favorite artist perform on stage is quite different from listening to their music at home or anywhere else being digitally played back. What makes a live performance so different is obviously the presence of the artist itself, but not just that. The unexpected and seemingly arbitrary improvisations on stage, the lack of total digital perfection and the moments when the artist becomes one with their music reflected by the expressions or gestures, make the whole experience so enjoyable.
I am not a person well versed with electronic music. If I’d heard some of the music I listen to now a few years back I’d have thought it rubbish or plain weird. Attending on-campus concerts and listening to selected pieces in a music class, I gradually grew to appreciate it. Both- being bored to hell and wishing the ear blasting awesomeness would never end, have been a part of the process. But the performances that had the most impact were always ones with the artist being on stage, playing with the elements, sometimes being a part of them.
In “Human Bodies, Computer Music“, Bob Ostertag poses an interesting question- that of incorporating the body of an artist working with electronic music, in the performance itself. Nine years back maybe this was a more difficult problem then it is today. If there was ever a solution to a seamless interface between man and machine then it’d have to be now and it’d have to be computer vision, which although does come with its own set of various problems, is still advanced enough for constrained scenarios. That and other media processing, sensor advancements make it easier than ever to incorporate an artist’s body in their performance. Pamela Z is an example of innovative media artists who use techniques varying from proximity sensing to gesture recognition in their performance and it definitely has a unique, more engrossing effect on the audience.
I worked on a project few months back where the user could interact with virtual objects around him (visible on the screen) to play different frequencies of sine tones. It’s called Virtual Synth, and I was told it had a good performative characteristic. I didn’t fully understand what this meant till I read the article above. If I imagine a fully developed, 10 times more robust, n-th iteration of this software I see an application with virtual instruments around an artist who could control all the parameters on screen which they once had to with keyboard and mouse, merely with hand movements and gestures, to create computer music while being part of the performance.
In the end though, it all comes down to what an artist has in mind for the audience to explore in their piece. I’ve been to shows where the artist would perform on-stage along with live visuals even, to concerts where there is virtually no performative aspect except controlling parameters on a laptop’s screen, but both being equally amazing. That’s the best thing about electronic music- the immense amount of control and flexibility it has that open up a wide range of possibilities to explore and unique ways of carving a niche in.