Hi,
I'm developing a C++ program that times itself, and needs to run on
Windows (yeah, I know) as well as Linux. Now, this isn't a big problem,
but it makes me curious. I use the clock() function from time.h to get
the time at the beginning of my simulation and at the end, and then use
the difference to measure how long the simulation ran. I'm noticing a
difference in behavior between Windows and Linux.
In windows, I will end up getting how much real time the simulation
took (i.e., if I suspend the machine for half a day and then let the
simulation finish, that half a day would be in the result). However, in
Linux, it seems to give me how much processor time the simulation took.
Can anybody explain why the same function from the same header file
behaves so differently?
Randy
--
TriLUG mailing list : http://www.trilug.org/mailman/listinfo/trilug
TriLUG Organizational FAQ : http://trilug.org/faq/
TriLUG Member Services FAQ : http://members.trilug.org/services_faq/