Can you tell me after how many cycles you see this error when you start Simics from phase_000? You can check the cycles with the “ptime” command in SImics.
-- Cansu On 06 Jul 2014, at 16:52, J Ahn <[email protected]<mailto:[email protected]>> wrote: >From the provided CloudSuite images, I am creating some additional phases of >CloudStone application. The simulated system produces the ZFS messages as follows. (My host machine does not use Solaris with ZFS system.) ------------------------------------------------------------------------------------------------------------------------------------------------ $] /opt/virtutech/simics-3.0.31/bin/simics -fast /home/flexus/workloads/cloudstone/2cpu/baseline/phase_000/simics/phase _000 -no-win Checking out a license... done: academic license. +----------------+ Copyright 1998-2007 by Virtutech, All Rights Reserved | Virtutech | Version: Simics 3.0.31 | Simics | Build: 1406 Host: amd64-linux +----------------+ www.simics.com<http://www.simics.com/> "Virtutech" and "Simics" are trademarks of Virtutech AB Use of this software is subject to appropriate license. Type 'copyright' for details on copyright. Type 'help help' for info on the on-line documentation. simics> run-command-file create_phases.simics simics> c SUNW-MSG-ID: FMD-8000-11, TYPE: Defect, VER: 1, SEVERITY: Minor EVENT-TIME: Sat May 19 03:13:22 GMT 2012 PLATFORM: SUNW,Sun-Fire, CSN: -, HOSTNAME: besim SOURCE: zfs-diagnosis, REV: 1.0 EVENT-ID: 6eac8d7b-7d99-6aaf-c063-cdbea30a0002 DESC: A Solaris Fault Manager component generated a diagnosis for which no message summary exists. Refer to http://sun.com/msg/FMD-8000-11 fo$ more information. AUTO-RESPONSE: The diagnosis has been saved in the fault log for examination by Sun. IMPACT: The fault log will need to be manually examined using fmdump(1M) in order to determine if any human response is required. REC-ACTION: Use fmdump -v -u <EVENT-ID> to view the diagnosis result. Run pkgchk -n SUNWfmd to ensure that fault management software is insta$led properly. [server_hme0 info] We don't own next Rx descriptor [server_hme0 info] We don't own next Rx descriptor [server_hme0 info] We don't own next Rx descriptor ------------------------------------------------------------------------------------------------------------------------------------------------ 2014-07-06 23:47 GMT+09:00 Cansu Kaynak <[email protected]<mailto:[email protected]>>: Dear Jeongseob, Can you please copy and paste the error messages? And please indicate if you see the messages in Simics or on your host machine. -- Cansu On 06 Jul 2014, at 16:35, J Ahn <[email protected]<mailto:[email protected]>> wrote: I am also wondering whether the ZFS diagnosis message affect the simulated systems or not. When creating some additional phases from the released phase-0, I often see the ZFS diagnosis messages in many applications(Cassandra, Streaming, Nutch, and etc) of CloudSuite. Does anyone who experience the similar problems ? 2014-07-06 19:14 GMT+09:00 J Ahn <[email protected]<mailto:[email protected]>>: Hi all, When creating the phases for Cassandra with 16cpus, the java exceptions occurs as follows. ------------------------------------------------------------------------------------------------------------------ simics> c [39/85] 37 0 0 404440 3141776 0 67 0 0 0 0 0 0 0 66 90 112043 253435 104683 91 8 1 78 0 0 404352 3129384 0 37 0 0 0 0 0 0 0 66 68 112976 256015 105540 91 8 1 SUNW-MSG-ID: FMD-8000-11, TYPE: Defect, VER: 1, SEVERITY: Minor EVENT-TIME: Thu Mar 1 03:40:13 GMT 2012 PLATFORM: SUNW,Sun-Fire, CSN: -, HOSTNAME: server SOURCE: zfs-diagnosis, REV: 1.0 EVENT-ID: e72c81aa-5c28-6b5f-9a35-886022834a70 DESC: A Solaris Fault Manager component generated a diagnosis for which no message summary exists. Refer to http://sun.com/msg/FMD-8000-11 fo$ more information. AUTO-RESPONSE: The diagnosis has been saved in the fault log for examination by Sun. IMPACT: The fault log will need to be manually examined using fmdump(1M) in order to determine if any human response is required. REC-ACTION: Use fmdump -v -u <EVENT-ID> to view the diagnosis result. Run pkgchk -n SUNWfmd to ensure that fault management software is insta$ led properly. log4j:ERROR Failed to flush writer, java.io.IOException: Bad exchange descriptor at java.io.FileOutputStream.writeBytes(Native Method) at java.io.FileOutputStream.write(FileOutputStream.java:260) at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:202) at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:272) at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:276) at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:122) at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:212) at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59) at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324) at org.apache.log4j.RollingFileAppender.subAppend(RollingFileAppender.java:276) at org.apache.log4j.WriterAppender.append(WriterAppender.java:162) at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251) at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66) at org.apache.log4j.Category.callAppenders(Category.java:206) at org.apache.log4j.Category.forcedLog(Category.java:391) at org.apache.log4j.Category.log(Category.java:856) at org.slf4j.impl.Log4jLoggerAdapter.info<http://org.slf4j.impl.log4jloggeradapter.info/>(Log4jLoggerAdapter.java:305) at org.apache.cassandra.db.Memtable.writeSortedContents(Memtable.java:164) at org.apache.cassandra.db.Memtable.access$000(Memtable.java:51) at org.apache.cassandra.db.Memtable$1.runMayThrow(Memtable.java:176) at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:30) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:619) 56 0 0 400560 3104880 28 154 0 0 0 0 0 59 0 89 53 108956 236869 107265 88 10 1 [server_sn0 info] Dropping packet to 127.0.0.16 - no route. 64 0 0 400048 3083632 35 143 0 0 0 0 0 79 0 74 54 115209 245856 109447 91 7 1 120 sec: 1354816 operations; 12048.5 current ops/sec; [UPDATE AverageLatency(ms)=1.71] [READ AverageLatency(ms)=23.3] 41 0 0 407048 3073368 93 575 0 0 0 0 0 184 0 48 45 111381 247377 105177 91 8 1 Feb 29 22:40:16 server svc.startd[9]: system/fmd:default failed: transitioned to maintenance (see 'svcs -xv' for details) 33 0 0 406352 3035792 40 348 0 0 0 0 0 59 0 71 65 108814 240641 100231 91 7 2 0 0 0 406344 3035784 0 23 0 0 0 0 0 0 0 63 55 114383 246439 104711 91 8 1 67 0 0 417304 3022168 0 55 0 0 0 0 0 0 0 62 91 110965 237690 105908 91 8 2 60 0 0 416184 3000560 0 43 0 0 0 0 0 2 0 65 65 112764 253245 107545 90 8 2 46 0 0 414992 2896872 0 163 0 0 0 0 0 0 0 83 68 109318 272641 104447 90 8 2 ------------------------------------------------------------------------------------------------------------------ I did not modify the images provided with EPFL at all. Could you experience the similar problem? Best regards, Jeongseob
