Interface Fractures is a series of audiovisual explorations Luka Prinčič has been developing in collaboration with the Slovenian Cinematheque since 2013. The series’ cinema-sound episodes share the same method and format: an immersive situation in the darkness of the cinema, the use of digital and open source tools for generating image and sound, a dichotomy between fixed composition and improvisation in time, a tendency for abstraction out of which fragments of the concrete arise, and playing with the synchronisation of sound and image. Although the creation process includes an examination of the contemporary human condition, conventional narration is not so important in the created performances and remains in the background. Interface Fractures is thus a fragmented excursion into abstract sound and the moving picture. Their synthetic integration emerges from the process of searching for fractures and subjectivities in a seeming impenetrability of polished and polarised interfaces – mediated, inter-machinic, interhuman.
Processing is a Java application and has had troubles in my Awesome setup for a while now in various ways. With some early versions (2.x) the main IDE window didn’t want to redraw following a resize. Now with new IDE in 3.x series this problem is gone, but now the main output window thinks it has a window decoration and is offset in a very ugly way, showing a wide grey bar at the bottom and slighly less wide one on the right (lack of top window decoration and a scrollbar probably.
These problems can be solved with a program called wmname before starting a java application.
The writing below used to be a synopsis for a some kind of vlog progress report about how the Turns Me On video was made. To be honest, I recorded the talk but didn’t like it, but since it actually detailed in a brief way the process I converted it into a blog post.
The idea of this post is to give a peek at the behind the scenes ideas and mechanics that lead to the way the Turns Me On video turned out.
At the initial brainstorming the basic problem in that early stage seemed to be this question: does a video for a track about sexual excitement really needs hypersexualized images of women in any way? I really wanted to turn this stereotype on its head somehow and first had the idea to go to clothes shops with a friend, a young male actor, and take skirts, high heels, sexy tights and other women’s clothes, go to changing rooms and have another friend shoot him trying things on. But I was already getting late with finishing the main track, not to mention I wanted to make remixes and so on. So the video was on hold all this time as I was working on music and I felt I need a more flexible plan. And, you know, the need for flexibilty usually translates into kinda do-it-yourself quick guerilla work. At least with me.
The intro does not leave you hanging: here comes an electro beat, lined with a relentless bassline and supported by a mix-up of samples from a well-known house track and a rhythmic litany by a male voice.
The title track has no refrain, no breakdown, just two long verses that keep on going and going. At first, the lyrics seem to be about those almost “stereotypically sexy” female adornments. The assumption that they are exclaimed with possible sexual aims is soon affirmed, but the lyrics of the second part of the track subverts the arousing into other daily phenomena not so typical of sexual hints.
My desire to talk about a fairly contemporary state of male sexuality – some kind of media programming of a “reptilian” part of the male brain to be sexually triggered with an image – is not so new. Subjective history takes me back to the solo performance called “fiberoptikal” where I wrote a song with verses: “frozen little images / are blocking my sight / making me uptight / they’re like bondage rope”, referring to the sexualized images of female body, including arousing erotic lingerie, specific body parts and their shapes, nail polish, makeup, high heels, specific body poses and movements etc. It was further developed in the performance “Frozen Images” by Wanda & Nova deViator (with Maja Delak), but didn’t make it onto our debut album “Pacification”, so only those that attended the “Frozen Images” concert performance(s) might know about it.
On the other hand, Emanat institute started to develop a new type of feminist and movement-centered burlesque performance. Informed by remix+edit appropriations via electronic dance music and contemporary technological development it was called “Image Snatchers”, a techno-burlesque. In autumn 2014 a workshop with the great performer Ursula Martinez took place, thanks to collaboration with the Ljubljana’s festival City of Women, where the seeds of another take on ‘programming of the male mind’ were planted: a burlesque number called “Turns Me On”. While researching for that piece I subjectively gravitated to a bit of personal history, a popular early house track with a moaning female voice, in my current view another example of voiced embodiment of hyper-sexualized desire, sonic enactment of submission and objectification, in a retrospect a metaphor of manipulation for profit and greed by white male capitalist.
While risking consequences of copyright infringement and a fall to a bootleg obscurity the reference to French Kiss by legendary Lil’ Louis is kept in the clear ear-view – on purpose. It should come as obvious within the context explained above, that this is a quotation, a reference to ponder about and critically reflect the contemporary exploitation of our bodies via frozen, arrested two-dimensional (sound-)images which make us turned on, to which we masturbate, to which we may climax.
Dedicated to all who suffer from erectile dysfunction.
stream/download at Bandcamp:
released August 15, 2016
written, produced and mixed by Nova deViator
vocals by Crucial Pink
mastering by Fred Miller
media & technical support: Radio Študent, Ljubljana
Multitouch support on Linux seems much simpler with a different library than SMT that I was unsuccessfully tryin last year. “Simple Touch” actually “just works” via evdev library. The only thing that was needed was to supply the right event device from /dev/input and also make it readable and writable for normal user.
This opens the door for writing multitouch interfaces in Processing, which however means a lot more work. For now it seems a good way to go with writing control GUI in SuperCollider, but eventually, the Processing possibility seems like very interesting and inviting.
So a final complete working example for my machine is this:
With the summer-time, a working-time on my new audio-visual piece, Interface Fractures III, begun. It is now almost confirmed that the date of premiere showing at Slovenian Cinemateque (Slovenska Kinoteka) is most probably 15/september. Since the plan was that we spend some quality sun&salt time at Croatian coast I brough some machinery with me to vacation. It’s always fun to work in the summer heat!
Anyway, with this next episode in the series I want to upgrade technically a “little bit”, so I acquired a better graphic card (Nvidia GTX960) and a multi-touch screen monitor with fullHD 1080p resolution. Adding also a 120GB SSD drive I needed to reinstall operating system (UbuntuStudio 14.4.1), separately compiled drivers for Nvidia and the rest worked pretty much out of the box (after some apt-get-ing). Multi-touch is application-dependent and my idea (for many years now) is to write custom interfaces for live sound/music/noise and visual composition and improvisation.
Here’s a little sketch in Processing that does the following: loads an image, takes a horizontal and vertical 1px slice, multiplies each slice into an image off-screen, and blends the two images together and displays the original and blended one side by side. Each frame this is calculated dynamicaly, the slices are determined by the position of the mouse.
Note: the image must be in the folder where your sketch is saved and it must be in dimension of 300×300 pixels.
I still need to test this in a fullHD/1080p situation. I wonder if the CPU can take it at 60 frames per second. I actually suspect not. So many pixels and not on the GPU.
* code by firstname.lastname@example.org
// load an image (should be inside the sketch forlder)
// get mouse data
// get slices from original image and draw two images off-screen
// get color/pixels data
// draw first image of slices off-screen
// draw second image of slices off-screen
// display original image
// draw transparent lines where the slices are taken
// blend the hor+ver images and display
// see http://processing.org/reference/blend_.html
Learning things has always been possible through two main routes (many other are possible though, surely): learning through a reference and learning by example. Personally it’s quite hard for me to learn through reference – it’s like learning grammar and syntax of a foreigh language without speaking the language. There always must be examples of use. Many of them. But who made examples? Others. So, learning from others in an open (free software) world is crucial element of today’s information society IMHO. In the old days I learned HTML from other webpages.
Here are two sites packed with different processing examples to learn from:
Luka Prinčič / Nova deViator: a musician, sound & media artist, engineer and dj. My sound goes from elektrofunk breakbeats to noise, drone and experiments. I’m one half of Wanda & Nova deViator, I run Kamizdat label and work at Emanat institute. I’m passionate about free software, cyberpunk, sci-fi, social awareness, critical expressions and peculiarity of contemporary human condition.
Like what you hear, see, read? Making music and art takes many hours of hard work. Consider helping me by becoming member at Patreon, paying for some free music at Bandcamp, donating bitcoin or simply via PayPal.me!