How to have a person move an object that is generated/projected by Isadora
-
downloaded simpleKinect and am going to look into it. Not able to find NI-Mate v.1. The project has potentially changed a bit. In meeting with the choreographer this was shown to me as the latest idea. It seems canned? Possibly using tracking but curious all of your opinions as it is the evolution of my initial question. Trying to learn what Izzy is capable of before going too far down rabbit holes.
Look at the 2:20 mark specifically of this [video](https://www.youtube.com/watch?v=-wVq41Bi2yE) -
Nice effect. Doubt it's canned if AEF were involved...
I'll have a ponder.At the moment my mind's heading off to MagicMusicVisuals if I'm honest, converting the OSC to MIDI to use as a deformer on an Interactive GLSL Shader patch. Finding / authoring a shader patch to give the visual effect you want however would be the domain of... well... a GLSL Shader authorer type d00d. Beyond me... I just swipe ones freely available online. (Is Izzy capable of this yet or do we have a timeline for it @Skulpture?) -
Marci...you have sufficiently lost me. I do not know what you are talking about. Sorry. GLSL Shader authorer type d00d. I have heard of open GL, understand basically what a shader is after looking it up. Wonder if Max/Msp/Jitter might be useful here. How does Izzy interact with Max?
-
Just thought I would wade in with a historical reference that might just trigger some other lines of thought. (admitting i need to go back thoroughly through the thread)
In the early 1990's I used an Amiga based system called Mandala. It was based on an 8bit video input card 'Live'. The Mandala software achieved very fast and reliable collision detection by allowing one to assign certain of the bit planes to the incoming video signal and others to graphics in ones environment. When different bitplanes collided in the same pixel space a range of functions could be triggered. it included gravity and attachment to left and right X coordinates. Surprisingly useful, a performer could grab looping anims move them around then throw them off with a flick of the wrist. There are many videos of performance and installations i did back in the day on my site.One of the first things I did when I came to Isadora in 2009 was try unsuccessfully to recreate facsimiles of some of that work of mine. I use Kinect, Leap, InfusionIcubeX and a bunch of other sensing devices. Like vanakaru i am not overly fond of kinect and find video camera solutions more reliable out of the studio.I think there would be a strong place for a collision detection actor as described by Marci. I too dream a dream! -
I agree. What has always baffled me is this:
There are ways of using javascript and other advanced calculations to detect two flat edges and when they it each other they bounce off each other. this can be done using sprites and envalope generators and inside range floats, etc.But with a kinect the edges are of course not just flat; they move and change quite rapidly - as well as having many edges. No idea where to even start with that! -
Collision detection actor sounds like it is needed and would be helpful to many but that implementation would be a challenge.
-
"But with a kinect the edges are of course not just flat; they move and change quite rapidly - as well as having many edges. No idea where to even start with that! "
What are we talking here... the edges of what? What is interacting with what in which direction? I'm only talking using the OSC Skeleton points. I guess you're talking about something in some way interacting with NI-Mate's ghost output/body outline rather than just the simple hand / foot / head co-ords etc?NB: I'm certainly no advocate of Kinect in any massive way - I just happen to have some kicking about to experiment with easily in front of the TV hence working down that route. -
PS: Izzy Interacting with Max... OSC. Think these days pretty much everything talks OSC.
@timeg - what's your website address? Curious to see...! -
@Marci I was talking about the depth image from a kinect. SO the multiple edges are the "outline" of a person.
-
Wow, I got to this thread late in the game.
3D collision detection should be pretty easy in Isadora as long as you have good data.I would be tempted to use 2 IR cameras. One in front and one above.XY from the front, and Z from above. In anycase adding the Z to the Quad detection should be easy.I have only played with rects and circles for detection,but if you want to determine more complex relations I would suggest looking into 'Point in Polygon' detection. see: https://github.com/substack/point-in-polygonSince the number of points in a polygon may vary, it becomes important to pass data sets together (unless you are defining a specific form). It will help keep your patch clean. see: http://dusxproductions.com/blog/pro-tip-single-patchcords-multiple-values/I would think that you could define a dynamic polygon based on the skeleton data that would be close to the actual figure.Then using this 'Point in Polygon' method you should be able to make the collision detections required.Again I would 'fake' the 3D to some extent by make X number of Z regions (just lowering the resolution in this dimension really)Its great to see that my blog has been a help :) -
Some context to my idea/dream
https://www.youtube.com/watch?v=jKB0d9vsfgA -
And here again a unified coordinate system either normalised or pixel based would make comparative calculation and position in Isadora one million times easier.
-
I've found an old patch that @gavspav made for me when talking about this a while ago.
I can't get it to work but it may be handy for some. -
@Fred It's in the feature requests. When you say normalized, what do you mean? I think the precentage measure isadora uses can do the job. It's just knowing which elements are based on stage width vs height that makes calculations difficult. Personally I like the scalability of this method. Using a little care you can make projects that adapt to different displays very easily.
-
DusX and Skulpture,
How to track an infrared LED in Isadora?. I can see an infrared light using Processing and simple open NI tracking IR; it is being picked up by the Kinect. Cannot figure out how to either see and use the IR LED to be tracked in Isadora and then used as a trigger. What would that actor configuration look like? Currently I'm simulating moving particles on the X, Y planes with a Mouse Watcher but would like to replace it with the IR LED as both DusX and Vanakuru suggested. -
infrared camera -> video in watcher -> sprite (to calibrate size and positioning) -> eyes -> gives you an x and y for the infrared dot.
-
Wii remote and OSC, I used years ago, sure it still dose the job.
best ...rainbow -
i don´t quite get your concept. :( the dancer needs to carry the Wii ? -
ofcourse you need a IR led as well for the actor/dancer ;))
-
gapworks- IR led on the dancer. Kinect watches the IR led.
Thanks dbini! I will check out.