Hi guys, I'm a newbie in tinyos. I read the FTSP paper and i tested FTSP
following the readme to verify the accuracy. I tested the algorithm on 10
motes and on the other two motes  I used BaseStation and RadiocountToLeds
just like described in the readme and then I followed the readme.matlab. The
length of experiment was 1 hour and the the average absolute error was 0.321
jiffy, and the maximum absolute error was 1.5 jiffy. Now, if a jiffy is 30.5
us how can be possible that the error is in the order of micro second when
in the algorithm the used timer is in millisecond?  Is it a stupid question?



--
View this message in context: 
http://tinyos-help.10906.n7.nabble.com/Why-microsecond-accuracy-in-FTSP-with-telosb-tp23558.html
Sent from the TinyOS - Help mailing list archive at Nabble.com.
_______________________________________________
Tinyos-help mailing list
[email protected]
https://www.millennium.berkeley.edu/cgi-bin/mailman/listinfo/tinyos-help

Reply via email to