Hey guys,

I did some test yestarday and I'm blown away by the results!

Latency from the suit heavily reduced and enhanced experience, this suit is
really surprising me with its results :D

Enjoy the videos

Lightroom Test Scene
<https://www.youtube.com/watch?v=uC2SZv3Z7ZE&feature=gp-n-y&google_comment_id=z121zljgiqyqzpf0d22tznzbjwnicpy4104>

Dark Souls Style Optimized Scene
<https://www.youtube.com/watch?v=Frd88EXVQEo>

Cheers

Nicolas

2016-08-06 9:58 GMT+02:00 Enter Reality <[email protected]>:

> Hey Adam,
>
> Thanks for the tips on Xsens
>
> I did some further tests and I was blown away by the results...basically
> I'm using the same smartphone I'm using as a VR device ( S7 Edge ) as a
> router wifi using 4g, and the latency between the real movement and the VR
> movement has been reduced by a lot! It's like being connected with the USB
> cable and the feedback is truly amazing!
>
> So instead of carrying around my tower pc, the router and so on currently
> I can bring just the laptop and that's it!
>
> I'm quite sure that, in order to demonstrate the lag, a First Person
> Shooter demo can be done just for playtesting, but what comes next will be
> the real challenge for me ( NDA at the moment, but I'll probably share
> something during the next couple of weeks )
>
> Perception Neuron does not support multi-level and the setup is a bit time
> consuming, but overall I'm very satisfied with the results...also
> considering that this is a side project I'm developing and I was able to
> have a wireless experience without using expensive tech or optical tracking.
>
> Nicolas
>
> 2016-08-05 23:17 GMT+02:00 Adam Sale <[email protected]>:
>
>> Hi again Nicolas. I haven't had the chance to test on a fast desktop yet,
>> was running the suit off of a Surface Pro 3 for an overseas job I just had.
>>
>> There was some slight delay in dealing with MVN studio and Motionbuilder
>> realtime. Again, I want to test with the vive on a higher end machine to be
>> able to truly comment. I don't think the latency was as bad as what I saw
>> in your video, bit it was still there. I will be back at work in the next
>> couple of weeks and do some more tests then. Will post back in this thread.
>>
>> One thing I like with the XSens, is how quick it is to setup. Doesn't
>> lose calibration easily, unless there is significant magnetic interference.
>> There is a debug mode where you can find the 'zones' where the tracking
>> breaks down.
>>
>> The multi level feature is one of the best features of the XSens suit.
>> I've been able to do climb three level homes with the laptop in the
>> basement, and still get exceptional tracking. Using the terrain around us
>> is such a big plus with these kinds of suits.
>>
>> Adam
>>
>> On Thu, Aug 4, 2016 at 11:35 PM, Enter Reality <[email protected]> wrote:
>>
>>> Hi Adam!
>>>
>>> Based on your tests with the Xsens suit, is the streaming of the mocap
>>> data almost realtime or there is a delay similar to the one you see in my
>>> video?
>>> I know that they work using similar IMU-based sensors, but since the
>>> price difference between them is alsmot 1 to 10 I was thinking to do the
>>> upgrade in a couple of months, but if I have the same delay I'll stick with
>>> the one I own :)
>>>
>>> 2016-08-04 23:49 GMT+02:00 Adam Sale <[email protected]>:
>>>
>>>> That's very cool, Nicolas. We've been building up our hardware assets
>>>> for just this kind of thing. I'm excited to test out Xsens and the Vive
>>>> with Unity or Unreal. Thanks for sharing!
>>>>
>>>> Adam
>>>>
>>>> On Thu, Aug 4, 2016 at 1:47 PM, Fabricio Chamon <[email protected]>
>>>> wrote:
>>>>
>>>>> ok thanks a lot for the deep explanation Nicolas.. All I need is a
>>>>> quick way to get acceptable mocap results to feed into my own crowd
>>>>> workflow. >From your description it seems that after some trial and error
>>>>> one is able to get it going and produce some decent animations. For the
>>>>> price maybe it's worth a try,
>>>>> thanks again.
>>>>>
>>>>> 2016-08-04 11:19 GMT-03:00 Enter Reality <[email protected]>:
>>>>>
>>>>>> Hi Fabricio,
>>>>>>
>>>>>> If you take a look at the forum there are quite a lot of people
>>>>>> complaining about lots of stuff, but it's also true that with 1500$ you
>>>>>> won't get a super-professional product, even if I can say that the money 
>>>>>> I
>>>>>> spent were well spent ( I backed the kickstarter campaign and got 2 of 
>>>>>> them
>>>>>> for a total of 1000$ )
>>>>>>
>>>>>> In general the suit works pretty good and I have been working with it
>>>>>> since January...in order to speed up the recording and to put the 
>>>>>> animation
>>>>>> onto a character I created my own pipeline which right now allows me to
>>>>>> record and process the animations in very short time.
>>>>>>
>>>>>> The time consuming part is at the beginning when you're start testing
>>>>>> the entire setup ( lots of things can go wrong and will probably go 
>>>>>> wrong )
>>>>>> The calibration itself ( if followed properly ) gives good results,
>>>>>> but the actual "virtual puppet" which is rapresented isn't a 1:1 copy of
>>>>>> you, since you can choose only the height of the character, nothing else.
>>>>>> Since it's a IMU based mocap suit there are issues when you try to
>>>>>> clap or you two-hand a sword for example, but this is something which can
>>>>>> be easily adjusted later inside Maya/MoBu.
>>>>>>
>>>>>> You can use two methods to record animation: via USB or via Wifi
>>>>>> USB works every time and it takes 1 second to be connected, while
>>>>>> Wifi connection is ( most of the time ) completely random, so sometimes
>>>>>> you're connected in 10 seconds while sometimes it takes up to 5
>>>>>> minutes...the developers themself are not sure why this is happening, 
>>>>>> which
>>>>>> is great :D
>>>>>>
>>>>>> Once you're connected the only thing you need to take care of is to
>>>>>> be away from magnetic source ( mobile phones, pc, and so on ).
>>>>>> The software used to record the animation allows for some tweaking (
>>>>>> hips fixed in place, smoothing and so on ) and in general it works quite
>>>>>> good, also considering that there are dedicated plugins for UE4, Unity,
>>>>>> Motion Builder and iClone.
>>>>>> Since January the sensors got magnetized last week, but that
>>>>>> happended because I constantly worked with my mobile phone in my hands or
>>>>>> very close to them, so I'm experiencing some drifting while 
>>>>>> recording...to
>>>>>> demagnetize you have to re-calibrate the sensors, which is a bit time
>>>>>> consuming, but in 20 minutes you're done.
>>>>>>
>>>>>> Overall I think that 1500$ for this suit are worthed if you don't
>>>>>> need a super-high mocap suit, this is a very good solution for hobbyst 
>>>>>> but
>>>>>> also for those who wants to create their own animation rather then paying
>>>>>> someone else and probably giving away the amount of 4 suits.
>>>>>> The tech issues are there, but it also depends on your setup.
>>>>>> With my tower pc the wifi streaming works great, with my laptop the
>>>>>> suit is not recognized, so if I need to record long animations in a large
>>>>>> space I need to move my tower pc ( and everything else ) outside, which 
>>>>>> is
>>>>>> a bit annoying but doable.
>>>>>>
>>>>>> Nicolas
>>>>>>
>>>>>> 2016-08-04 15:48 GMT+02:00 Fabricio Chamon <[email protected]>:
>>>>>>
>>>>>>> Hey Nicolas, very nice! I'm particularly interested in the mocap
>>>>>>> suit. Does it work well? How is the setup/calibration time? Any 
>>>>>>> pros/cons?
>>>>>>> I can see in the perception neuron website that you have to be
>>>>>>> extremely carefull with your calibration or the system really will not 
>>>>>>> work
>>>>>>> good enough. How is your experience with that?
>>>>>>>
>>>>>>> Thanks for sharing, nice experiment!
>>>>>>>
>>>>>>>
>>>>>>> Em quinta-feira, 4 de agosto de 2016, Enter Reality <
>>>>>>> [email protected]> escreveu:
>>>>>>>
>>>>>>>> Yep, I couldn't find a good enough sword so I put that in :D
>>>>>>>>
>>>>>>>> Possibly during the weekend I'll updated the scene with a
>>>>>>>> multi-level environment ( which helps a lot with the ridiculous 
>>>>>>>> restriction
>>>>>>>> on mobile ) and maybe turn it into a game, but I doubt it ( single 
>>>>>>>> person
>>>>>>>> development with zero programming knowledge ), but I'm curious to see 
>>>>>>>> how
>>>>>>>> other people will use this setup
>>>>>>>>
>>>>>>>> 2016-08-04 13:00 GMT+02:00 Sebastien Sterling <
>>>>>>>> [email protected]>:
>>>>>>>>
>>>>>>>>> Is that Artoria's sword :P
>>>>>>>>>
>>>>>>>>> On 4 August 2016 at 12:39, Olivier Jeannel <[email protected]
>>>>>>>>> > wrote:
>>>>>>>>>
>>>>>>>>>> Super ! Please keep us posted, very inspirating !
>>>>>>>>>>
>>>>>>>>>> On Thu, Aug 4, 2016 at 11:54 AM, Enter Reality <[email protected]>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi guys,
>>>>>>>>>>>
>>>>>>>>>>> I just want to share one of the projects I'm developing using
>>>>>>>>>>> the GearVR and a mocap suit.
>>>>>>>>>>>
>>>>>>>>>>> In short I'm walking inside a VR environment using the mocap
>>>>>>>>>>> suit to walk around and interact with the environment
>>>>>>>>>>>
>>>>>>>>>>> GearVR Test Demo <https://www.youtube.com/watch?v=-4IOA9VjxKc>
>>>>>>>>>>>
>>>>>>>>>>> I'm using a Samsung Galaxy S7 Edge and the Noitom Perception
>>>>>>>>>>> Neuron Suit, and all the data is sent to the smartphone via 
>>>>>>>>>>> Wifi...a bit
>>>>>>>>>>> laggy, but considenring the setup I'm quite satisfied by the 
>>>>>>>>>>> results.
>>>>>>>>>>>
>>>>>>>>>>> I've shared the entire project on the UE4 forum
>>>>>>>>>>> <https://forums.unrealengine.com/showthread.php?119165-Perception-Neuron-Dark-Souls(ish)-Test-Demo-Full-Body-VR-(-Downloadable-project!-)>
>>>>>>>>>>> if someone want to test the scene
>>>>>>>>>>>
>>>>>>>>>>> The video shows just a very simple environment, but future
>>>>>>>>>>> updates will include more content and features.
>>>>>>>>>>>
>>>>>>>>>>> I'm not planning to focus on games using this setup, but I guess
>>>>>>>>>>> that this is a good way to show everyone what I'm doing.
>>>>>>>>>>>
>>>>>>>>>>> Best Regards
>>>>>>>>>>>
>>>>>>>>>>> Nicolas Esposito
>>>>>>>>>>>
>>>>>>>>>>> ------
>>>>>>>>>>> Softimage Mailing List.
>>>>>>>>>>> To unsubscribe, send a mail to [email protected]
>>>>>>>>>>> odesk.com with "unsubscribe" in the subject, and reply to
>>>>>>>>>>> confirm.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> ------
>>>>>>>>>> Softimage Mailing List.
>>>>>>>>>> To unsubscribe, send a mail to [email protected]
>>>>>>>>>> odesk.com with "unsubscribe" in the subject, and reply to
>>>>>>>>>> confirm.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> ------
>>>>>>>>> Softimage Mailing List.
>>>>>>>>> To unsubscribe, send a mail to [email protected]
>>>>>>>>> odesk.com with "unsubscribe" in the subject, and reply to confirm.
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>> ------
>>>>>>> Softimage Mailing List.
>>>>>>> To unsubscribe, send a mail to [email protected]
>>>>>>> odesk.com with "unsubscribe" in the subject, and reply to confirm.
>>>>>>>
>>>>>>
>>>>>>
>>>>>> ------
>>>>>> Softimage Mailing List.
>>>>>> To unsubscribe, send a mail to [email protected]
>>>>>> odesk.com with "unsubscribe" in the subject, and reply to confirm.
>>>>>>
>>>>>
>>>>>
>>>>> ------
>>>>> Softimage Mailing List.
>>>>> To unsubscribe, send a mail to [email protected]
>>>>> with "unsubscribe" in the subject, and reply to confirm.
>>>>>
>>>>
>>>>
>>>> ------
>>>> Softimage Mailing List.
>>>> To unsubscribe, send a mail to [email protected]
>>>> with "unsubscribe" in the subject, and reply to confirm.
>>>>
>>>
>>>
>>> ------
>>> Softimage Mailing List.
>>> To unsubscribe, send a mail to [email protected]
>>> with "unsubscribe" in the subject, and reply to confirm.
>>>
>>
>>
>> ------
>> Softimage Mailing List.
>> To unsubscribe, send a mail to [email protected]
>> with "unsubscribe" in the subject, and reply to confirm.
>>
>
>
------
Softimage Mailing List.
To unsubscribe, send a mail to [email protected] with 
"unsubscribe" in the subject, and reply to confirm.

Reply via email to