Archive | July, 2015

Still: Intrinsic Gravity

14 Jul

PC demo, 1st place at Ghetto Scene 2014. Original file.

Pasarinho – Rainer Scheurenbrand

13 Jul

Rainer Scheurenbrand with a fabulous music video straight from the forest.

“It’s Alright (Baby’s Coming Back)” by Eurythmics

12 Jul

The 80s are an interesting period for psychedelic aesthetics. Hippies were declared outdated by punk rock in the 70s and after that psychedelic aesthetics had to blend with the images of the time to stay relevant. And they did — with great success. A lot of the visual tropes of psychedelic aesthetics had already established themselves in visual culture. Straight lines and plain surfaces now cooled down the arabesque imagination of the 60s, yet still shapes constantly morphed into each other and lines began dancing out of nowhere. What used to be associated with the influence of mind altering substances now got linked to limitless possibilities of new technology and media.

Mondial TM

11 Jul

Always There – Cachopou (2015)

11 Jul

 

Fresh frames from Yoshihide Sodeoka’s latest music video by Spain’s Cachopou. Bring on the distortion balls!

As Seen On TV

11 Jul

There are videos that I know got some time on MTV and others I’m not sure about. Truthfully, I didn’t watch much of their programming when I came to learn that music videos weren’t actually much of a priority. I can think of a handful that I’d like you guys to see though. Maybe you saw them on MTV first, and if you enjoyed them, I’ll be the first to thank the network for bringing them to the people. If you haven’t seen them or even one of them, I’m glad the DPV has provided.

I, for one, somehow managed to mostly miss out on the Gorillaz when the project was at its most popular. Years later, I glanced over a friend’s shoulder as he watched this video.

I had to know what was going on. After watching all their other videos and reading up a bit, I’ve got to admit that I haven’t really found any clear reasons for anything that’s going on, but I think it’s cool as hell regardless. Thank you, Grey, for showing me this one.

Once again, I find myself amazed by simplicity, and this video astounds me. A great deal of editing or design is often involved in the creation of psychedelic art, but in this case it’s just raw performance. Never in my life will I be able to pull off a dance like this.

One friend of mine told me he distinctly remembered the morning he saw this video for the first time. He was just about to leave for school, but ended up being late because he just had to stay and watch this masterpiece to the end.

The chances that you haven’t heard the song are flat out nonexistent. Mmmmaybe, just maybe though, this video will today reach a new set of eyeballs out there for the first time.

Nothing groundbreaking going on here, but damn if it’s not catchy. This one is just a guilty pleasure.

And finally, a caricature of all the ills of mankind. No subtlety here.

I hope at least one caught your fancy, whether for the first time or for the first time in a while.

Russel Brand’s The Trews

10 Jul

Trews, the YouTube channel by English comedian, actor, author, activist and celebrity Russel Brand, promises to “unravel the matrix of modern media and reveal the gleaming reality beyond connecting us all to each other through pure consciousness. Or it’s true news. Trews.”

Similarly to Jason Silva, Brand brings YouTube viewers big, sparkling ideas in concise and accessible form. However, with a different mixture of elements: Less Californian techno-utopian ideas, and more social- and media critique on the state of global society; less scenic imagery and more hilarious humor.

The trews YouTube channel tackles issues like “Why so hard on the silk road?”  “Are we all terrorists now?”  and “Can Jeb Bush finally win WW2 for the Nazis?”. Above one of Brand’s more lucid videos, on the question of God’s existence.

 

 

Semi-Permanent 2015 Opening Titles

9 Jul

Journey through the layers of the mind // Memo Akten

8 Jul

A visualisation of what’s happening inside the mind of an artificial neural network.

In non-technical speak:

An artificial neural network can be thought of as analogous to a brain (immensely, immensely, immensely simplified. nothing like a brain really). It consists of layers of neurons and connections between neurons. Information is stored in this network as ‘weights’ (strengths) of connections between neurons. Low layers (i.e. closer to the input, e.g. ‘eyes’) store (and recognise) low level abstract features (corners, edges, orientations etc.) and higher layers store (and recognise) higher level features. This is analogous to how information is stored in the mammalian cerebral cortex (e.g. our brain).

Here a neural network has been ‘trained’ on millions of images – i.e. the images have been fed into the network, and the network has ‘learnt’ about them (establishes weights / strengths for each neuron).

Then when the network is fed a new unknown image (e.g. me), it tries to make sense of (i.e. recognise) this new image in context of what it already knows, i.e. what it’s already been trained on.

This can be thought of as asking the network “Based on what you’ve seen / what you know, what do you think this is?”, and is analogous to you recognising objects in clouds or ink / rorschach tests etc.

The effect is further exaggerated by encouraging the algorithm to generate an image of what it ‘thinks’ it is seeing, and feeding that image back into the input. Then it’s asked to reevaluate, creating a positive feedback loop, reinforcing the biased misinterpretation.

This is like asking you to draw what you think you see in the clouds, and then asking you to look at your drawing and draw what you think you are seeing in your drawing etc,

That last sentence was actually not fully accurate. It would be accurate, if instead of asking you to draw what you think you saw in the clouds, we scanned your brain, looked at a particular group of neurons, reconstructed an image based on the firing patterns of those neurons, based on the in-between representational states in your brain, and gave *that* image to you to look at. Then you would try to make sense of (i.e. recognise) *that* image, and the whole process will be repeated.

We aren’t actually asking the system what it thinks the image is, we’re extracting the image from somewhere inside the network. From any one of the layers. Since different layers store different levels of abstraction and detail, picking different layers to generate the ‘internal picture’ hi-lights different features.

All based on the google research by Alexander Mordvintsev, Software Engineer, Christopher Olah, Software Engineering Intern and Mike Tyka, Software Engineer

Hattie Stewart – Oxymoron (Damien Hirst Projections)

7 Jul

Andy Baker animated Hattie’s illustrations.