masterxeon1001 writes:
I'm still working with this little black box of doom and after much headache and help I got it in the BGE. I just wanted to showcase some of the progress made with it.
masterxeon1001 writes:
I'm still working with this little black box of doom and after much headache and help I got it in the BGE. I just wanted to showcase some of the progress made with it.
13 Comments
Yes! Yeeeeeeees! Blender needs tools like this to make it more productive. Blender is very powerful but interaction with meshes is often akward. This work presents exciting possibilities. I hope it represents the future of the Blender interface. I cant wait to play with it. More videos please!
Hope it can work with Kinect one day.
Its great that someone is trying to integrate Leap into Blender, but I wouldn't get my hopes up about this person's code making it to the trunk. The author calls it a "hack and slash" method. Thats not to say I am unimpressed. Having tasted blender coding/scripting myself, I know what a challenge it is to create something like this.
XD smashed autodesk!
We should make a plan of what's the best way to integrate LEAP so that one wouldn't have to even touch the keyboard for switching between different tasks. My thoughts:
1. Two hands should by default mean navigating, you could push, pull, stretch, and rotate the viewport easily in your hands. You could navigate with one hand too if no objects are selected.
2. One hand should by default mean object interaction. You would first select an object by tapping on it, and then gestures should activate transform functions. If you wish to grab for example, you could just shake the object a few times and it the grab tool would activate. Rotate for rotating, pinch for scaling. You could confirm or cancel the action by showing thumbs up or thumbs down with the other hand. You could navigate around the object any time by just using two hands, no interruptions.
3. In sculpt mode one could select any of the existing brushes or select a special mold clay brush that would work like real life molding.
4. You could also use both hands for object interaction or sculpting, but first you would have to toggle two hand navigating off by showing a gesture, for example both hands closed as fists.
Oh, and
5. There should be a mode for hands-on like interaction for collision objects too (like in this video). But this should be a separate thing imo.
i could really use someone with python know how. if anyone is wanting to explain to a newb some stuff about python. drop me an email at [email protected]
Once again, for this technology to take off, eg to use while sculpting in Blender, we need a real time stereo 3D display feature. can use gsls but as you can see when trying to grab objects in Unity demo, you need 3D displays to get real world interaction.
Warning: At this stage it would be unwise to pre-order.
If the device actually exists as advertised, then there should be no problem in purchasing the it when it is scheduled for released on May 13.
If you chose to ignore this advice, then $79.99 might be worth a valuable life lesson for you.
its very much like the ads. I assure you.
With "0.01 mm precision"?
Does it display "a quick visual of the data that's been captured."
http://www.youtube.com/watch?v=vukBCXLG_u0
This could be great for hand/puppeteering motion capture. Anima Vitae made a series ( https://www.youtube.com/watch?v=bNpT2u4cHkk&list=UUm0b5xPLnJzQPMnU2TiXVcg&index=10 ), using a custom built hand motion capture solution. Some of the arm movement is scripted/phsyics, and some hand animation was added on top, but it did apparently increase the amount of animation each animator could do each week manyfold.