Hi Steve,
Thanks for the information! Good to know that a new version with better source control is going to be released :-) Yeah I'll be happed to try the new version and keep up with your pace. So what's the job for a tester? Jiayuan _____ From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Steve Reinhardt Sent: Tuesday, August 14, 2007 10:16 AM To: M5 users mailing list Subject: Re: [m5-users] infinite cache access with atomic_add Yes, that's the latest official release, but we've fixed some bugs since then and have recently revamped the cache coherence significantly. I'm guessing that you are running into a bug that may be fixed with our latest code. (Or if it's still an outstanding bug, we'd want to fix it in our latest code anyway.) We've recently converted to Mercurial internally (http://www.selenic.com/mercurial) and are getting geared up to provide public access so that people can more easily keep up with changes... would you be willing to be a beta tester for this? Steve On 8/13/07, Jiayuan Meng <[EMAIL PROTECTED]> wrote: I forgot to mention. I am using M5v2.0 beta3. I suppose this is the most recent release? Thanks. Jiayuan Hi Jiayuan, There may be some bugs with the load locked/store conditional code. What version of m5 are you using? What configuration are you using (cache levels, coherence protocol, timng vs. atomic mode, etc.)? Steve On 8/11/07, Jiayuan Meng <[EMAIL PROTECTED]> wrote: Hi all, I encounter an infinte loop when running a multi-threaded program on a multicore simulator with a control processor and 4 slave cores. This atomic_add, as printed in the trace, seems to keep accessing the cache forever and I have to aborted the simulation manually. Any clues on it? Thanks a lot! Jiayuan 6 685000250000: system.slaves0 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+12 : stl_c r1,0(r16) : MemWrite : D=0x00000 7 685000252000: system.slaves0 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+16 : beq r1,0x1200e8144 : IntAlu : 8 685000254000: system.slaves1 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+8 : addl r1,r17,r1 : IntAlu : D=0x00000000 9 685000244000: system.master T1 : @_ZN9__gnu_cxx18__exchange_and_addEPVii+4 : ldl_l r0,0(r16) : MemRead : D=0x00 10 685000255000: system.slaves1 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+12 : stl_c r1,0(r16) : MemWrite : D=0x00000 11 685000257000: system.slaves1 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+16 : beq r1,0x1200e8144 : IntAlu : 12 685000245000: system.slaves3 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+4 : ldl_l r1,0(r16) : MemRead : D=0x0000000 13 685000257000: system.master T1 : @_ZN9__gnu_cxx18__exchange_and_addEPVii+8 : addl r0,r17,r1 : IntAlu : D=0x000 14 685000258000: system.slaves3 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+8 : addl r1,r17,r1 : IntAlu : D=0x00000000 15 685000258000: system.master T1 : @_ZN9__gnu_cxx18__exchange_and_addEPVii+12 : stl_c r1,0(r16) : MemWrite : D=0x 16 685000260000: system.master T1 : @_ZN9__gnu_cxx18__exchange_and_addEPVii+16 : beq r1,0x1200e8164 : IntAlu : 17 685000259000: system.slaves3 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+12 : stl_c r1,0(r16) : MemWrite : D=0x00000 18 685000261000: system.slaves3 T0 : @_ZN9__gnu_cxx12__atomic_addEPVii+16 : beq r1,0x1200e8144 : IntAlu : _______________________________________________ m5-users mailing list m5-users@m5sim.org http://m5sim.org/cgi-bin/mailman/listinfo/m5-users _____ _______________________________________________ m5-users mailing list m5-users@m5sim.org http://m5sim.org/cgi-bin/mailman/listinfo/m5-users _______________________________________________ m5-users mailing list m5-users@m5sim.org http://m5sim.org/cgi-bin/mailman/listinfo/m5-users
_______________________________________________ m5-users mailing list m5-users@m5sim.org http://m5sim.org/cgi-bin/mailman/listinfo/m5-users