Date/time: 6/February 2015
Location: Goldsmiths, University of London, Cinema
On 6/February, we presented projects developed during the Hackathon on Interactive Computer-Generated Audiovisuals (6-7 December 2014) and subsequent weeks. The projects are open-source, and can be downloaded from https://github.com/avuis.
About the projects:
In ABP, a PureData patch creates music and sends data to the Cinder visuals software via OSC. Several visual parameters like color, alpha, zoom, repetition of objects, tempo are given by the PureData patch.
Authors: Adam John Williams, Bruce Lane, Piotr Nierobisz
Project link: https://github.com/AVUIs/AdamBrucePiotr
drawSynth consists of a GUI to control sound and image. Users can draw shapes and select colours. By doing this, users control the synthesis engine. Each pair of vertices of a shape will control an FM oscillator – one vertex for carrier and the other for the modulator. The project is built with openFrameworks for graphics and interaction, and Max/MSP for sound, using FM synthesis. OSC is used for communication between both.
Authors: Antonio Creo, Diego Fagundes, George Haworth
Project link: https://github.com/AVUIs/drawSynth
Esoterion Universe consists of an empty 3D space that can be filled with planet-like audiovisual objects. The objects can be manipulated and can be given different appearances and sounds. Users can navigate in space and the audiovisual outcome is influenced by that navigation. Generic, media neutral terms such as warmth, sharpness, size and oscillation are used to characterise and connect sound and visuals. A semantic approach was chosen for this connection instead of a one-to-one parameter mapping. The GUI consists of sliders distributed concentrically, in the shape of a star graph, embedded in the centre of the object. It integrates aesthetically with the objects. openFrameworks with openGL is used for graphics and interaction, and Max/MSP for sound. OSC is used for communication.
Authors: Borut Kumperscak, Coralie Diatkine, Matthias Moos, Will Gallia
Project link: https://github.com/AVUIs/EsoterionUniverse
GS.avi is a gestural instrument that generates continuous spatial visualizations and music from the input of a performer. The features extracted form a performer’s gesture defines the color, position, form and orientation of a 3-demensional Delaunay mesh — its composite triangles, vertices, edges and walk. The music, composed using granular synthesis, is generated from features extracted from the mesh — its colors, strokes, position, orientation and patterns. The project was created using Processing and MAX/MSP. OSC is used to communicate between the two.
Authors: Ireti Olowe, Miguel Ortiz, Will Brown
Project link: https://github.com/AVUIs/GS.avi
Modulant allows for the creation of images and their sonification. The present implementation is built upon image-importing and freehand-drawing modules that may be used to create arbitrary visual scenes, with more constrained functional and typographical modules in development. The audio engine is inspired by a 1940’s synthesizer, the ANS, that scans across images. In this scanning, one axis is time and the other axis is frequency. Modulant thus becomes a graphical space to be explored sonically and vice-versa. The project is built with Processing for graphics and interaction, and Ruby with Puredata for sound.
Authors: Berkan Eskikaya, Louis Pilfold, Alessandro Zippilli
Project link: https://github.com/AVUIs/Modulant
The performance is co-organized by EAVI (Embodied AudioVisual Interaction) group at Goldsmiths, University of London, in collaboration with London Video Hackspace . It is part of the Enabling AVUIs research project being conducted at EAVI, and supported by the European Union (Marie Curie programme). More information: http://avuis.goldsmithsdigital.com