I originally wanted to due a music generative project based off of possibility and an input from the mic. But after researching online and especially finding out about the music group Autechre I changed my mind. I mainly got my inspiration from their patches. Sound designs were learnt both through the youtube DeliciousMaxTutorials and http://sounddesignwithmax.blogspot.com/. Reference for the reverb subpatch: taken from https://cycling74.com/forums/reverb-in-max-msp.
Here is a recording sample of the piece being played:
Code as follows:
]]>First off, a wav file of single piano notes played consecutively is analyzed. While Max randomly selects portions of the wav file to play in snippets, the frequency of the audio being played is analyzed and triggers the 1st higher-pitched drones in intervals. Meanwhile, the 2nd drone patch can be triggered by using the keyboard as a midi-keyboard.
The drone is achieved via subtractive synthesis. The pink noise generator is send through filters, only letting pass certain frequency bands. Then, the subtractive synthesis is done with a handful of inline reson~ objects.
The ‘analyzer~’ object is referenced from the maxobject.com website.
Delay is added to all sound effects. Piano melody can also go through a noise gate at will. The speed of the piano sampling can also be manipulated, which will immediately also effect the speed of the self-generated higher pitched drones.
Here is an example of the music being played:
Code is Here:
]]>I created a noise reduction subpatch. Furthermore, a degrading fft subpatch is also used and linked to the output, playing along with the other signal. After these two subpatches, the audio signal is then run through the original patch where it is stretched out in real time (slowed down). This is done by using the delay effect.
I also added another simple visual presentation, that is very similar to the one we made in class. A japanese city pop song was used in demonstration, to achieve that ‘vaporwave aesthetic’.
Another demo: https://soundcloud.com/thewx/assignment4-demo/s-K8iMc
Top level patch:
Noise Reduction Patch:
Degrading Patch
Visual Patch:
]]>An inspiration source I am having for this project is as follows:
This is a lot more advanced than what I’m hoping to achieve, but definitely includes certain artistic styles that I’d like to imitate in my project as well.
]]>Here are the results:
Kalimba Original IR1 IR2 IR3 IR4 Kalimba IR1 Kalimba IR2 Kalimba IR3 Kalimba IR4I really liked the sound of this, and decided to make a track through Ableton Live.
Track explanation as follows:
Kalimbas:
Apart from a tiny bit of compression (Ableton built-in) and limiter (George Yohng’s W1) added, not much else mixing is done (my apologies).
]]>]]>
I started with a simple image of a bird-eye view picture of an island. I ran my image through the system 630 times to be exact. The effect or changes can be quite noticeable at the beginning, but they slowly become harder to notice as the signal gets twisted even more. I also noticed that the algorithm itself has a few bugs, leaving some blocks of pixels unsorted at all. Especially in the middle (middle of islands) where brightness levels are even and high, the effect won’t be present no matter how many times the process is ran.
For more info about pixel sorting, please see: http://datamoshing.com/2016/06/16/how-to-glitch-images-using-pixel-sorting/
]]>