• Products
    • Isadora
    • Get It
    • ADD-ONS
    • IzzyCast
    • Get It
  • Forum
  • Help
  • Werkstatt
  • Newsletter
  • Impressum
  • Dsgvo
  • Press
  • Isadora
  • Get It
  • ADD-ONS
  • IzzyCast
  • Get It
  • Press
  • Dsgvo
  • Impressum

Navigation

    • Register
    • Login
    • Search
    • Categories
    • Recent
    • Popular
    • Tags

    Gear porn

    Hardware
    7
    16
    6351
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Fred
      Fred last edited by

      "Traditional" motion capture systems use multiple cameras pointing into a space. They can track markers as individual points but better they can track arrays of markers as rigid bodies. With rigid bodies you cannot mistake one marker for another as they are unique. To be correctly identified a marker (part of a rigid body- they use 3 markers at least per rigid body) has to be seen by at least 3 cameras. From this the system can triangulate the exact position (mm accuracy) of the marker, in the case of a rigid body it means giving precise position and orientation (something you cannot do with normal tracking or even a kinect). We have 2 motion capture systems both from Natural point (the cheapest option for motion capture but not so cheap) the one I mentioned here has 8 wide angle high resolution (gige with poe on cat6 lines) cameras that are put around the space, the system is calibrated and then we tell it where we want the zero point to be (in the centre of the room). This system run autonomously from our other rigs and sends a live TCP stream of the position and orientation of all our rigid bodies. You can also use suits with 30 markers to track humans in space at very high speed and accuracy. This is what is used to capture the motion of dancers to apply it to 3D models in hollywood animations. Here is a cheesy sample of the system at a trade show 

      https://www.youtube.com/watch?v=s2oIBUO4mY8
      More info on the system we use is here
      http://www.naturalpoint.com/optitrack/products/motive/
      Here is some of the things we have been doing with it
      http://vimeo.com/61531402
      We made this so we can draw augmented reality live (using a camera as the input source for drawing) 

      http://www.fredrodrigues.net/
      https://github.com/fred-dev
      OSX 13.6.4 (22G513) MBP 2019 16" 2.3 GHz 8-Core i9, Radeon Pro 5500M 8 GB, 32g RAM
      Windows 10 7700K, GTX 1080ti, 32g RAM, 2tb raided SSD

      1 Reply Last reply Reply Quote 0
      • bruper
        bruper last edited by

        jaw dropping

        17"MBP 2.93GHZ Core2Duo mid 2009 - OSX10.11.6 - 8GB, 1TBCrucial_SSD, izzy 3.0.7

        1 Reply Last reply Reply Quote 0
        • Skulpture
          Skulpture Izzy Guru last edited by

          I'm not even jealous or anything.... ;-)

          Graham Thorne | www.grahamthorne.co.uk
          RIG 1: Custom-built PC: Windows 11. Ryzen 7 7700X, RTX3080, 32G DDR5 RAM. 2 x m.2.
          RIG 2: Laptop Dell G15: Windows 11, Intel i9 12th Gen. RTX3070ti, 16G RAM (DDR5), 2 x NVME M.2 SSD.
          RIG 3: Apple Laptop: rMBP i7, 8gig RAM 256 SSD, HD, OS X 10.12.12

          1 Reply Last reply Reply Quote 0
          • eight
            eight last edited by

            @Fred Thanks for the info. The bottom video is very cool and funny.

            --8

            Analysis: http://post.scriptum.ru | Synthesis: http://onewaytheater.us
            Twitter: https://twitter.com/eight_io | Flickr: http://www.flickr.com/photos/eight_io/
            Github: https://github.com/eighteight | MulchCam: https//mulchcam.com
            MulchTune: https://itunes.apple.com/us/app/mulch-tune/id1070973465 | Augmented Theatre: https://augmentedtheatre.com

            1 Reply Last reply Reply Quote 0
            • Aolis
              Aolis last edited by

              Ooooo, Ooooo I have some questions:

              Are you using Isadora at all in this set up?

              If you are, how are you connecting your MoCap data with Isadora? How do you map the rotations of your MoCap data to your Isadora actors?
              If you are not using Isadora, then what software are you using to integrate your live video feeds with MoCap data?
              Kind regards,
              Alex

              Media Artist & Teacher
              MacBook Pro, Apple M3 Max, 128 GB
              Sonoma 14.3.1

              1 Reply Last reply Reply Quote 0
              • Fred
                Fred last edited by

                No, we dont use Isadora, we make everything in openframeworks. This stuff is not possible in Isadora. I have used Izzy with the mocap as I wrote some software to get the position data via OSC. We do the video overlay via hardware (using the ATEM switcher from blackmagic and generating a live alpha mask out one of the HDMI outputs in real time. The ATEm can do the blending in HD with less than a frame delay. I rarely use Izzy these days but sometimes give workshops on it. However my hardware is Izzy compatible and I worked a long time to get very high quality image workflows so I shared it here.

                Fred

                http://www.fredrodrigues.net/
                https://github.com/fred-dev
                OSX 13.6.4 (22G513) MBP 2019 16" 2.3 GHz 8-Core i9, Radeon Pro 5500M 8 GB, 32g RAM
                Windows 10 7700K, GTX 1080ti, 32g RAM, 2tb raided SSD

                1 Reply Last reply Reply Quote 0
                • Aolis
                  Aolis last edited by

                  Fred,

                  I love what you are doing and thank you for explaining the video overlay signal path.  
                  I am currently trying to integrate a VICON system with Isadora and am having some strange problems calculating rotation data from the position data I get via OSC. Thus my question about how you might map rotations. 
                  Basically I am getting XYZ data and am trying to calculate the rotations based on comparisons between the 5-6 points I am tracking on a prop.  
                  At the moment I am struggling with a strange kind of flipping that occurs when I try rotating a prop more then 180 degrees or so. . . . 

                  Media Artist & Teacher
                  MacBook Pro, Apple M3 Max, 128 GB
                  Sonoma 14.3.1

                  1 Reply Last reply Reply Quote 0
                  • Fred
                    Fred last edited by

                    When you get data from the system we use only quaternions, they provide the correct 3d location and rotation information. Without being able to use quaternions and rotation matrices getting anything useful from a full motion capture system is pretty much impossible. I tried to do this kind of precise work with Isadora but found too many inconsistencies and limitations in Isadora's 3d environment. Quartz composer maybe of more use for you as it has a more advanced set of 3d tools, but honestly without rotation matrices and view ports and other standard 3d tools you will be very limited. What exactly are you trying to do with rotation? Maybe I can get you started with some code?

                    http://www.fredrodrigues.net/
                    https://github.com/fred-dev
                    OSX 13.6.4 (22G513) MBP 2019 16" 2.3 GHz 8-Core i9, Radeon Pro 5500M 8 GB, 32g RAM
                    Windows 10 7700K, GTX 1080ti, 32g RAM, 2tb raided SSD

                    1 Reply Last reply Reply Quote 0
                    • Aolis
                      Aolis last edited by

                      Wow, Thank you!
                      I have been attempting to map a 3D asset to a physical object via VICON>OSC>ISADORA. 
                      One possible goal in this current research is to produce a 1:1 relatoinship with properly distorted video onto a physical object that is being manipulated in realtime. 
                      Quartz may well be the way to go as I am at least superficially familiar with it. I am unfamiliar with what it offers in the way of 3D tools and interaction via OSC. I will need to spend some time researching the 3D tools available in/for Quartz Composer. . . .
                      I've gotten to the point where Isdaora can see what I think is the raw XYZ information produced from VICON. Each point VICON tracks sends ts XYZ coordinates on a something like a -5000 to +5000 mm 3 axis matrix. 
                      I have been mucking about with math and trying to use the Calc Angle 3D actor to map steady rotation between a digital prop and a physical prop. At the moment the best I have breen able to do is get the 3D asset to follow a very limited range of motion. My current calculation causes the digital asset to flip and spin when I cross certain thresholds of rotation. I'm guessing this might be the kind of problems you ran into?
                      I had a colleague suggest today that I am dealing with "Gimble Lock." Which "is when two rotation axes overlap." They also mentioned that I should investigate "quaternions." . . .  
                      This is a bit of a false task though. In response to your quesiton "What exactly are you trying to do?" - Exactly, I am trying to create a curriculum for a new class that leverages Isadora in conjunction with VICON. We will also be looking at Quartz, MAX/Jitter, Motion Builder, and a number of other theatre centric performance media technologies such as DMX, LANBox, etc. 
                      What I am really doing is looking for is a good challenge AND finding a practical reason to integrate Isadora with a VICON system. Why Isadora? I am primarily teaching this class from the perspective of Isadora as a media design tool for rapid prototyping of interactive media systems - using it as a kind of glue for various media design techniques and technologies. Oh, and there is a VICON system in the EMMA lab I have the good luck to be teaching in.
                      Perhaps a simpler relationship between the XYZ data that does not include accurate rotations can still offer a theatrically viable interaction? . . . .
                      Thank you!

                      Media Artist & Teacher
                      MacBook Pro, Apple M3 Max, 128 GB
                      Sonoma 14.3.1

                      1 Reply Last reply Reply Quote 0
                      • Fred
                        Fred last edited by

                        The maths involved in calculating rotation from position is difficult and problematic (it is never exact and depending on the method you use you can end up with gimble lock). Isadora is not able to easily make these calculations. If you can switch to using real rotation data from the vicon (quarternions), maxmsp has some objects to handle them http://cycling74.com/forums/topic/ann-quaternion-objects-for-max/ Although dealing with this in any node based media tool will be pretty difficult and inefficient. You can also try look inside this tool, it is about as good as you can get for rotation from xyz position and it is quite limited (to reduce wildly wrong rotations). https://code.google.com/p/tryplex/ Somewhere in there is a halfway conversion from xyz to rotation. I know the folks who made it and this was as good as it could work. Using something like openframeworks is actually quite easy and i have been teaching it to young students and they pick it up quick. It is free, open source, cross platform, almost unlimited and much more powerful. The learning curve is no steeper than Isadora and it gives students especially an environment they can use well into the future. I wish I had found it earlier!

                        http://www.fredrodrigues.net/
                        https://github.com/fred-dev
                        OSX 13.6.4 (22G513) MBP 2019 16" 2.3 GHz 8-Core i9, Radeon Pro 5500M 8 GB, 32g RAM
                        Windows 10 7700K, GTX 1080ti, 32g RAM, 2tb raided SSD

                        1 Reply Last reply Reply Quote 0
                        • P
                          PatrickPagano last edited by

                          hi, nice gear. We used Organic Motion and jMonkey engine for Networked MoCap with Dancers in Japan, NY, Florida and Vancouver sharing the same space. we used OSC for the bone rotations and collisions. This is a very impressive setup!

                          1 Reply Last reply Reply Quote 0
                          • P
                            PatrickPagano last edited by

                            PS as for motion builder! I ended up getting incomplete data streams and corrupted .BVHs so beware!

                            1 Reply Last reply Reply Quote 0
                            • Fred
                              Fred last edited by

                              I have used motion builder a little but not had these problems- is it a problem from the data stream from Organic Motion?

                              http://www.fredrodrigues.net/
                              https://github.com/fred-dev
                              OSX 13.6.4 (22G513) MBP 2019 16" 2.3 GHz 8-Core i9, Radeon Pro 5500M 8 GB, 32g RAM
                              Windows 10 7700K, GTX 1080ti, 32g RAM, 2tb raided SSD

                              1 Reply Last reply Reply Quote 0
                              • First post
                                Last post