GMU:I and my Max/Mojjo Krenz: Difference between revisions

From Medien Wiki
No edit summary
No edit summary
Line 1: Line 1:
Semster Project : Chicken Tracker
We humans communicate. At work, with friends, colleagues and with relatives. To classify the messages communicated to us, we use our senses. In direct verbal communication, our auditory perception is primarily engaged. Peripherally one could think the acoustic channel is clear and unmistakable. What may be simply true in individual cases, however, reveals itself in everyday life as a diverse, complex act. Human communication and perception is multi-layered. Besides the acoustic channel we use optical, olfactory as well as tactile stimuli to communicate with our fellow human being. What happens to the understanding of our counterparts when visual communication is restricted, feeling and smelling are even completely eliminated, and when eye contact, facial expressions and gestures are distorted by the Internet?

The current pandemic and the associated lockdown limit verbal, interpersonal exchange to two elementary channels. Acoustic communication and visual communication made possible by video telephony are coming to the fore in our everyday lives. However, contact via the medium of the Internet lacks the possibility of expressing empathy and feelings.  For this purpose I have invented an „Chicken Tracker". 
As a support tool to the usual video chat, the „Chicken Tracker" is supposed to visually amplify a feeling. 
The basic idea is simple. Participant A starts talking, initially appears completely realistic in the video chat - if A talks himself into a rage, the volume of his acoustic output increases steadily. The amplified signal in turn turns his skin color increasingly red through the emotion tracker - until A appears completely red. Once in a rage, the volume of A's output increases until it collapses and A turns into a chicken for the time in which the volume is higher than X. The whole idea is about loosening up in online discussions and not taking yourself too seriously.
The project has several stages of development.
First, I wanted to change certain colours in a Max video patch. Therefore I build a patch exchanging bright / dark colours by a matrix. (see SC)
"Greenscreen"  
"Greenscreen"  


Line 10: Line 31:
https://www.studioroosegaarde.net/info
https://www.studioroosegaarde.net/info
https://www.studioroosegaarde.net/project/space-waste-lab
https://www.studioroosegaarde.net/project/space-waste-lab
Semster Project.
In times of endless zoom/bbb/etc. meetings its easy to loose patience. Some might need to see visually themselves turning red. The idea of my project is to change the skin color of the user inside a zoom meeting. The parameters for changing the color is decided by the volume of speech. Basically you turn yourself red on screen - while screaming.
The project has different development stages.
1. changing specific skin colours into red colour  in max video using the volume input of the mic
2. changing specific skin colours into different shades of red colours - floating - using the volume input of the mic
3. connecting video of max msp to zoom - implementing the skin changing ability for both partys

Revision as of 16:00, 18 January 2021

Semster Project : Chicken Tracker

We humans communicate. At work, with friends, colleagues and with relatives. To classify the messages communicated to us, we use our senses. In direct verbal communication, our auditory perception is primarily engaged. Peripherally one could think the acoustic channel is clear and unmistakable. What may be simply true in individual cases, however, reveals itself in everyday life as a diverse, complex act. Human communication and perception is multi-layered. Besides the acoustic channel we use optical, olfactory as well as tactile stimuli to communicate with our fellow human being. What happens to the understanding of our counterparts when visual communication is restricted, feeling and smelling are even completely eliminated, and when eye contact, facial expressions and gestures are distorted by the Internet?



The current pandemic and the associated lockdown limit verbal, interpersonal exchange to two elementary channels. Acoustic communication and visual communication made possible by video telephony are coming to the fore in our everyday lives. However, contact via the medium of the Internet lacks the possibility of expressing empathy and feelings. For this purpose I have invented an „Chicken Tracker". 
As a support tool to the usual video chat, the „Chicken Tracker" is supposed to visually amplify a feeling. 
The basic idea is simple. Participant A starts talking, initially appears completely realistic in the video chat - if A talks himself into a rage, the volume of his acoustic output increases steadily. The amplified signal in turn turns his skin color increasingly red through the emotion tracker - until A appears completely red. Once in a rage, the volume of A's output increases until it collapses and A turns into a chicken for the time in which the volume is higher than X. The whole idea is about loosening up in online discussions and not taking yourself too seriously.


The project has several stages of development. First, I wanted to change certain colours in a Max video patch. Therefore I build a patch exchanging bright / dark colours by a matrix. (see SC)





"Greenscreen"

1. try 2 videos in 1 patch - goal will be they run mixed into each other

File:VIDEO_PLAY1.maxpat

https://www.studioroosegaarde.net/info https://www.studioroosegaarde.net/project/space-waste-lab