"Listen" live music
-
@juriaan said:
Mac has this, Windows version not..
Thanks @Juriaan, I should have said my description is for Mac. Even so, it is not a āsimpleā technique and the variation of sound frequencies and levels in music can make it even more complex to maintain a workable calibration.
However, it can be a very powerful effect that tightly integrates sound and moving image.
Best wishesĀ
Bonemap
-
Really interested in how you actually do it. Getting a note out of Isadora features is something that is quite complex to say the least..
-
@juriaan said:
how you actually do it
Hi @Juriaan,
The way you describe using max appears to be similar.
I isolate the sound of a musical instrument using a contact microphone or very close micāing, through a pre amp audio interface. In Isadora I use multiple instances of the Sound Frequency Watcher per audio channel and calibrate each to listen for a narrow band of frequencies to match the individual notes of the scales played by the instrument.Ā
I build this into a user actor with a series of calculators between each instance of the frequency Watcher. That way I can duplicate and move the calibration of the frequency ranges of many Watchers all at once and in series.
This has worked well for instruments like guitars and harps. The thing with an individual note is it is often played as a chord so it is usual for more than one frequency range to be active simultaneously.
I should add that I have not used this technique to replicate anything scientific. I have used it to make generative art, so I have not been too tight with the tolerances and tend to use triggers and parameter constraints based on ranges rather than individual notes. Having said that, I did initially build a prototype for individual notes but only in a single octave. After, experiencing the āpainā of doing that, I decided less accurate ranges was going to be more efficient to achieve.
Best wishes
BonemapĀ
-
@cristina_spelti is the audio material monophonic or not?
-
Hello,
I was just doing a workshop about Sound Analysis during Werkstatt 2017, here are some of my reflexions.
I agree the best way is to use Max/msp where you have the best tools dedicated to sound analysis and its very easy after that to send information through OSC to use it inside Isadora.
A better way is, possible if the musician/s use midi instrument, is to send directly midi information to Isadora to trigger what you want. It's a pain to make sound from midi to retransforming it to OSCā¦.
But if you want to stay inside Isadora, the last version (I dont know if its available because I use beta version, but 2.6 is coming soon) have a much better sound analysis, particularly a Sound Frequency Bands, working quite well to trigger sound.
Another trick to use sound inside Isadora is to use the routing facilities of your sound card, some (as RME does), are able to reroute output towards input, and there is also the solution to reroute output to input through a numeric cable, ideally ADAT with 8 channels.
Jacques
-
@jhoepffner said:
Sound Frequency Bands,
Thanks Jacques, the list of new and updated features for the 2.6 release will be interesting reading. Ā I am not sure what happened to soundflower on the Mac but internal routing of audio becomes a recurring issue, particularly for keeping the frequency analysis within Isadora. I will definitely explore your suggestions.
@cristina_spelti, your question about a āsimpleā way, I believe is to use Isadoraās internal sound analysis, unless you also have MaxMSP or Midi Instruments that can be networked with Isadora.
Best wishes
BonemapĀ
Ā
-
Hi all!
@bonemap @crystalhorizon @jhoepffner @JuriaanĀ
Thank you for this discussion. Actually I still have to choose how to set up my work. The source of the sound will be a string orchestra completed by an electronic drums (aFrame drum), a piano, a midi keyboard, live voices, 2 laptops for electronics sounds.
I would like to use only Isadora (his internal sound analysis) and maybe a midi connection with the keybord ... but for this I have to study a lot because I know very little about the use of midi.I'm still confused but this discussion is helping me to understand.
Grazie!
-
This is a good way but take too time. I think that I can't do this alone, without orchestra and I have no time to rehearsing. I have to find something effective but simple
Thanks
all my best!
-
Hi,Ā
Take a look aĀ imitone
-
Interesting software ! Great for kids š
-
I think SoundFlower is defunct now. I've heard of people using Sound Syphon ($40 USD) but haven't myself because of the pricetag (I don't need audio routing very often).
Best wishes,
Woland
-
Ā Another possible technique would be to send numerous frequency readings as OSC to wekinator (a computer learning tool).
There you could train Wekinator to recognize triggers... this may prove to be a rather effective way to 'recognize' notes.
Wekinator can then send a OSC message back to Isadora if a trigger/note is recognized.
I haven't tested this... but it should absolutely be possible.
I've done similar with video.
-
-
-
Hi!
Someone can explain to me wich actor is possible to use to listen the music "beat"?
My idea is - in 3d particles -Ā using beat to add obj.
Thanks
Cri
-
Hehehe, that is a trick one. There is no 'Watch for BPM' actor in Izzy.
But let's see or we can come up with something
1. Use an other program (like MAX MSP or PureData (free)) to get the BPM and send it using OSC to Izzy.
2. Calculate the BPM by using the general rule of thumb. 1 BPM = (reading of the Hertz) * 60 (Since it is not beats per second, but beats per minute)
Since we can't use Izzy Frequency watchers on a Sound Player I rather go with 1, since the second option causes a lot of headache that we have to fix with some JavaScript... (And since Izzy doesn't provide an Audio API we can't simple put filters on the values causing a lot of coding to make this possible using math..)
-
if you are using frequency analysis in Isadora, you can quickly use a frequency watcher connected to a 'tap tempo' to get the BPM.
Tap tempo is a very hand little actor to get the BPM or hz of any input.Ā
-
-
The tap tempo actor works well but you have to tap in manually. Linking it to a keyboard watcher (such as space-bar) works well.Ā
-
you can use a sort of workaround to use tap tempo and frequences. Assuming that the beat is marked by an base drum you can check the main frequence produced by it ( usuallyĀ around 80 100 hz ) then conect that frequence band to a comparator and this one to tap tempo. keep in minde that this setu works fine with simple beat, dance or house music, if you have a more complex sound it can retun false positive. also i suggest you to add some smother or similar actor after tap tempo 'couse is not so precise. I did some test some month ago and even if recive a midi clock some time it mark a bit different time.
here a quick example