Skip to content
A MaxForLive device AV0 developed to explore live human-computer collaboration in the field of audio visual, dedicated to anyone believing in computers as partners in the creative flow.
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
images
AV0.amxd
LICENSE.txt
Layout Debugger.amxd
Levels.amxd
README.md
actions.txt
call-scene.maxpat
icon_timer.png
main.js
method-selector.maxpat
setup-scene.maxpat
shapes.js
sustain-check.maxpat
timeline.maxpat
timeline.txt
timer.maxpat
titles.maxpat
types.txt
utilities.js

README.md

hero

AV0

AV0 is an exploration on live human-computer collaboration in the field of audio visual, dedicated to anyone believing in computers as partners in the creative flow.

The project was started in September 2016 with the initial desire to make a piece to tour to venues. During its development however, I realized I was introducing a new perspective on performing audio visuals, and that the device I was building could become an expandable framework on top of which one would build their own set. My plan is to keep building it, and my dream is to do this with anyone else wants to help.

About this documentation

This documentation provides a quick access to trying AV0 and a high level view of the code you'll find in this repository. I'm sure there are plenty of important things I forgot to include, so please send your feedback and questions. You can do so by:

  • Creating an issue on this repo
  • Emailing me ay hellotoluca [at] gmail [dot] com
  • DM me via Facebook

If you whish to learn more about the concept behind AV0 you can do so on my website.


Getting started

This project was developed as a tool for myself and it therefore still lacks the proper packaging that would make it immediate to integrate it in your set. So in order to get you started as fast as possible I decided the best route was to share a Live Set already setup where you should be able to see what it does instead of reading a list of instructions.

  1. Download and launch this Live Set template which included all you need to try it.
  2. Activate Live's playback which will tells the visuals's window to starts drawing.
  3. Play notes (from E2 to G3) or launch the Demo clip to see visuals and hear sounds.

And that should be enough to get you hooked a little, hopefully.

NOTE: if you don't hear sound but everything seems setup correctly on your end, try deleting AV0 device from the track and the undo the action. This is the only workaround I found to a known but hard to replicate issue in MaxForLive where MIDI In/Out sometimes gets blocked.


About the Live set

overview

In this section I'll go over the setup of the Live Set mentioned in the Getting Started section. The image above shows what you should see after launching it. That includes the a Live Set with one track called Demo Track and the visual's window called video-window. In the Demo Track's detail view, you should see one big device called AV0 Group where I grouped together the following devices.

AV0

Detail AV0

This is the main device responsible for generating and controlling visuals. For more information on this device see the section below called The AV0 Device.

Sounds

Detail Sounds

This is just a drum rack with some sounds. This device plus the AV0 device make the bare bones for connecting visuals to sounds. One thing to mention here is the presence of an operator

Effects

Detail Effects

Currently AV0 has 3 simple effects, such as Opacity, Decay and Audio Feed. By themselves these effect have only an impact on the visuals, so combining them with audio effects help bringing the whole thing to life. That's why all these devices are grouped inside the AV0 Group, so Macro controlers can be used to controll both visuals' and audio's effects.

Levels

Detail Levels

This device is a little exploration on feeding filtered audio signal values into visuals using send and receive objects. All it does is letting the audio to pass through while grabbing signal's valuse to send to the AV0 device. The Audio Feed parameter on the AV0 device controls the amount of that value being passed to visuals.

Mental model

Mental Model

Image above shows a diagram go how the different elements in the demo Live Set interact with each other. The AV0 device works in Live as a MIDI through so that MIDI messages can be passed to the Live instruments positioned afterwards. In the meantime it uses the same MIDI input to send commands to the JS object and generate the visuals.
The Levels device works as a throughput for audio, and it sends back collected values to AV0 to affect visuals.

Mind the BPM

Tempo

Before telling more about the devices in the set I'd like to point your attention on the BPM set at 220bpm. There are reasons based on my personal style I won't get into. But you should know that the lower the tempo the slower the visuals will be updated. At 4/4 tempo, 220bpm means that each beat is fired at a spead slighlty higher than 1/64th of a second. The AV0 device uses half of that speed with a metro set on 32n. So if you want to use this device at a slower BPM you would have to change that metro value inside the timer.maxpat file.


The AV0 device

Detail AV0

As you probalby noticed the AV0 device's UI is divided in several areas. In this section I'll explain what each area does, going from left to right.

1 • Made JS edits?

Compiler

This function allows you to recompile the JS code in case you made some changes. The file main.js, which has autowatch property set to 1 shouldn't need this action, but any change to all other files requires to compile the JS. The reason why I put this and the following section at the very left is to make them more easily accessible while developing the device, eventually they should not even be visible at all.

2 • Notes input

Notes Input

MAPPING MAPPING MAPPING MAPPING MAPPING PADSSSSSSSSSAH!! Used for tesigng and debugging. At a surface level this section's purpose is to show what pad is being pressed and with which velocity.

3 • Output setup

Output Setup

This is where you indicate the Output resolution of the screen or projector you are using, and the viewport aspect ratio you desire to cut out for your visuals. To see the viewport mask hit V key while either the window or the AV0 device are on focus and playback is on.

4 • Window

Window

This allows to set some of the window's properties. Useful when you're reharsaling and don't need or want to connect a second screen or you're preparing audio material and want to hide the visuals.

5 • Controls

Controls

This area is highly unexplored. Current available effects are Opacity, which sets element's transarency, Audio Feed which enables audio signal to impact on the grid's color intensity, Decay which sets the ease and lifespan of elements (0: snappy moves and sudden fade, 127: slow moves and long fading). All other controls are hooked up in Max and in the JS but they currently do nothing.

6 • Timer

Timer

This is the part that enables the fun and upredictable collaboration bethween human andcomputer. When activated, the AV0 device will link the countdown represented by the progress bar ad teh bottom of the window, to the live set and the visuals. Once the time is up, Timer will stop all playing sounds, it will generate a new sets of visuals and block incoming MIDI messages for 3 seconds. This will most likely leave you in your underwear on the stage, hence the icon.


Grids and layers

Visuals are arranged based on a simple grid system of rows and columns, as one would do in approaching a graphic design. The viewport where visuals are displayed is divided in a number of columns and rows. Each cell generated by this subdivision can allocate one visual element.

Grids

The cell defines where the element will be positioned, its maximum size and the roaming areas at its disposal.

Cells

There is a total of four layers each of which has a different grid subdivision and can therefore host different types of visuals, quantity and size.

Layers


Mapping

In this section I'll explain how MIDI notes are mapped to visuals.
My controller of choice for using Live is Ableton Push 2, which means that the design and development process was sort of naturally affected by this preference. Specifically in taking advantage of the 64 pads grid and limiting, although I used only few of them, my controls to 8 knobs. I'll therefore reference this controller, usign some illustration from Ableton, to explain how pads are mapped. The translation to any other MIDI controller should be straigth forward.

Push

The illustration above shows, on the left, Push's pads grid, and on the right, you can see the macro sections in which I divided it. If we take one of this sections (image below) you can see how each bank of 16 pads represents 4 layers (L1, L2, L3, L4) consisting in a row of pads; and 4 actions (A1, A2, A3, A4) assigned vertically for each column.

Pads

As you might have figured, if you read the part titled Grids and Layers each row of four pads launches actions for each of the four layers.
There are four actions for each layer and their behavior is currently very simple. However the JS and MaxForLive patches allows to expand of these actions and have the AV0 device choosing from a selection of them every time a new piece and set of visuals is generated. If you whish to expand these actions the three files you should look at are method-selector.maxpat which can select a random action for every piece, actions.txt which tells the device what are all actions available, and elements.js where methods are called. Lastly please keep in mind that no all pads are mapped but only those going from MIDI note 4 to 115, this allows for 7 banks of 16 pads each. MIDI note 127 is used to generate the high pitch sound used to signal the generation of a new piece. See image below for reference.

Notes


Files in this repository

Simply a list of all the files you'll find in this repository, and a brief description on what they do. Please note: give its very alpha stage most if not all of the following files might look like they are offering functionalities that are not used or some method might look as designed to be apparently uselessly scalable. The hope of course is to fullly elaborate all of them soon.

MaxForLive files

  • AV0.amxd — This is the main MaxForLive device.
  • Layout Debugger.amxd — This device is used to test out new elements and how they behave in different layouts.
  • Levels.amxd — This device sends levels values via a send object that are then received by AV0 to affect elements and such. It's not a clean solution but though.

MaxMSP files

  • call-scene.maxpat — This file is responsible to stopping all sounds, blocking temporairily the user input and calling a new scene by banging setup-scene.maxpat. It's used every time the timer reaches the end, when enabled.
  • method-selector.maxpat — This file is responsible for assigning an action to each MIDI note. To see how MIDI notes are mapped please see the Mapping section.
  • setup-scene.maxpat — This file is responsible for creating a new scene. Sets the grid's subdivision, decides how many elements to add to each layer, what tipe and size. A lot od randomness goes on here, one day I hope it'll be more intelligent.
  • sustain-check.maxpat — This file is responsible to tell the js methods wether at least one note in a row is still pressed or not. To learn what I mean by rows please see the Mapping section.
  • timer.maxpat — This file bangs the js object and controls the countdown based on Live global transport.
  • titles.maxpat — This file is responsible for showing titles every time a new scene is created.

Text files

  • types.txt — This file is a list of names referring to elements that can be created. It's loaded and by setup-scene.maxpat to select random elements from it and then generate them via the relative method inside characters.js file.
  • actions.txt — This file is a list of actions elements can do when there is incoming MIDI messages. It's used by method-selector.maxpat which, at the creation of every new scene, will pick a random action variation for each note. At the current stage of AV0 this approach is not necessary given to the limited set of available actions developed so far.

Javscript files

  • main.js — This is the main js which is primarili responsible for listening to user's input and drawing on the window.
  • elements.js — This is where elements are defined, generated, organized on the grid and where actions are defined. Among all things I whis it was solved better, this is one of the areas I believe needs the greater improvements to make this device flexible enought to be able to generate elements and animate them.
  • utilities.js — This is where basic setup and other utilities are.



Secret keys

When the window or the device are on focus.

  • Press V: to see the viewport within which visuals are projected. Useful when going fullscreen on unusual aspect ratios and need to calibrate the projector.
  • Press T: to see again titles after they disappear. Pressing notes again will make them fade away.
You can’t perform that action at this time.