bluemonday

Created With: Processing.js, Kinect, After Effects

Song: Blue Monday by New Order

I approached this video project as a visual music piece. I began by breaking the song into phrases and then deeper into instruments and polyrhythms. I wanted to bring the electronic dance atmosphere of the song to life by capturing a lot of body movement and dance moves. I played with lots of bright colors and experimented with layering effects and kinetic typography to add a psychedelic dream sequence feel to the video.

I created all of the visuals using an XBOX Kinect in real time and live recorded all of the movement with 3D motion capture that I coded making use of available open source libraries in Processing.js. I chose to experiment with the Kinect as I wanted to explore the collision of the abstract digital world with the photo-real world. I mapped each letter of the word "bluemonday" to the 3D point clouds being generated in real time through the Kinect to create a unique texture and typographical element throughout the piece. 

midi_me

Built With: Processing.js, Minim.js, Optical Flow

midi_me is an interactive experience that I built using Processing.js. It tracks the user's body movement and triggers different piano notes depending on where their motion is located. Instead of having to pick up and learn an instrument, you can use midi_me to generate music through body movement.

Check out my code repo on Github here :)

Interactive Music Videos

Built With: Processing.js, Minim.js

I programmed this series of interactive visuals in Processing.js, using the Minim.js library so that the visuals respond to any imported song as well as input from the user such as key-presses and mouse clicks. I am hoping to soon host these interactive programs on a website for public use with any song that the user inputs.

Experimental Video

Objective: To display how humans can use algorithms, systems and rules to produce aesthetically pleasing and chance-based images and results.

Medium: All graphics and animations coded in Processing.js, original code with some adapted code from “Generative Art” by Matt Pearson

 

Generative Art, programs compiled and layered in Final Cut Pro


Music: Euphoric, Podington Bear / CC

Objective: Exploring what collage could look like in motion and how layering and keying the graphics and photography affects the visual aesthetic.

Medium: All graphics created in Adobe Illustrator, animated in After Effects, edited in Final Cut Pro


Music: Data, Podington Bear / CC

Hologram Dance Pack

Materials: Transparent acrylic, iPads, cardboard

Created By: Jamie Haberman and Avni Barman

Dancers: Jet Kwok, Beau Foley, Celine Kiner and Alyssa Allen

Toiletunez

Components: Arduino Uno, Micro-SD Card Adapter, Micro SD Card, Speaker, Proximity Sensor

Built With: Processing.js, Arduino, PHP, mySQL

Team: Jamie, Naylee, Brianna

Toiletunes addresses that awkward moment we all face at school or at work... You're in a stall in the bathroom doing your thing and someone else comes into the stall next door and you both sit in awkward silence hoping you won't make an embarrassing sound. To eliminate this awkward moment, why not bring music into this space with a stall radio! Whenever you walk into the stall and sit down, Toiletunes detects your presence and starts playing. Toiletunes in turn leads to a lively bathroom environment and a great conversation starter!

Working in a group of three, my team developed this project based around concepts of telepresence, product design, service design and interaction design. We utilized a web based storage of data, Arduino microcontrollers for controlling physical components, and front-end visualization techniques to interface with the physical components of ToiletTunes.

 

Check out our GitHub Repo here and some demo videos below!

  • email
  • linkedin