GMU:Max and the World/Leon Giacomo Goltermann: Difference between revisions

From Medien Wiki
No edit summary
No edit summary
 
(11 intermediate revisions by 2 users not shown)
Line 1: Line 1:
[[:File:homework_1.maxpat]]


[[:File:homework2_leon.maxpat]]
[[File:Bildschirmfoto 2021-05-12 um 18.08.24.png|400px]]


Prototype to test my Installation ideas. (it is reduced to triggering sounds with buttons insted of tracking).


'''New version with instruction in the patch.  
I built a patch that uses computer-vision to control sound-plaback with head-gestures.
[[:File:prototyp instalationsinterface.maxpat]]
Here is the patch and a video that explains how to use it and what you can do with it.
'''


here you can find soundfiles to try the patch (they are not finished, just to try) or you use your own sound/max-library
[[:File:facetracking_sampler_leon_g.maxpat]]
https://cloud.uni-weimar.de/s/znnFMEzCJGEcYBD


the purpose is documented in the video https://cloud.uni-weimar.de/s/ig53DAm3naP7Rfz
{{#ev:youtube|mef2j1WzQoI}}




As an evaluated project for the this course i want to produce a prototype that shows the possibilities of the installation - Lesekomposition -  (wt) i am building.
It should provide the connection of sounds and words. The sounds need to be triggered (by knobs or with sensor data). It should also be possible that sounds are changed/manipulated by the combination of word/triggers; there should be a library of sounds for each trigger and different presets for selecting different lists.


[[:File:versuch sprache rythmus.maxpat]]
The patch was developed in the process of building an interactive installation experimenting with the extension of language in the combination of sound and language.
 
I realized the Prototype - compo - which provides an instrument to create future interactive compositions between words and music.
 
It is documented over here: https://wwws.uni-weimar.de/kunst-und-gestaltung/wiki/GMU:Artists_Lab_IV/Leon_Goltermann

Latest revision as of 15:12, 14 May 2021

Bildschirmfoto 2021-05-12 um 18.08.24.png


I built a patch that uses computer-vision to control sound-plaback with head-gestures. Here is the patch and a video that explains how to use it and what you can do with it.

File:facetracking_sampler_leon_g.maxpat


The patch was developed in the process of building an interactive installation experimenting with the extension of language in the combination of sound and language.

I realized the Prototype - compo - which provides an instrument to create future interactive compositions between words and music.

It is documented over here: https://wwws.uni-weimar.de/kunst-und-gestaltung/wiki/GMU:Artists_Lab_IV/Leon_Goltermann