---------- Forwarded message ----------
From: J Ahn <[email protected]>
Date: 2014-07-07 0:17 GMT+09:00
Subject: Re: Java exception occurs on cassandra-16cpu
To: Cansu Kaynak <[email protected]>


The error occurs at 120766000 cycles from phase_000.
I attached the log with ptime.

-----------------------------------------------------------------------------------------------------------------------------
$] /opt/virtutech/simics-3.0.31/bin/simics -fast
/home/flexus/workloads/cloudstone/2cpu/baseline/phase_000/simics/phase_000
-no-win
Checking out a license... done: academic license.

  +----------------+    Copyright 1998-2007 by Virtutech, All Rights
Reserved
  |   Virtutech    |    Version: Simics 3.0.31
  |     Simics     |    Build: 1406  Host: amd64-linux
  +----------------+
    www.simics.com      "Virtutech" and "Simics" are trademarks of
Virtutech AB

Use of this software is subject to appropriate license.
Type 'copyright' for details on copyright.
Type 'help help' for info on the on-line documentation.

simics> ptime
processor            steps         cycles  time [s]
server_cpu1  1168578842000  1168578842000   584.289
simics> run-command-file create_phases.simics
simics> c

SUNW-MSG-ID: FMD-8000-11, TYPE: Defect, VER: 1, SEVERITY: Minor
EVENT-TIME: Sat May 19 03:13:22 GMT 2012
PLATFORM: SUNW,Sun-Fire, CSN: -, HOSTNAME: besim
SOURCE: zfs-diagnosis, REV: 1.0
EVENT-ID: 6eac8d7b-7d99-6aaf-c063-cdbea30a0002
DESC: A Solaris Fault Manager component generated a diagnosis for which no
message summary exists.  Refer to http://sun.com/msg/FMD-8000-11 for more
information.
AUTO-RESPONSE: The diagnosis has been saved in the fault log for
examination by Sun.
IMPACT: The fault log will need to be manually examined using fmdump(1M) in
order to determine if any human response is required.
REC-ACTION: Use fmdump -v -u <EVENT-ID> to view the diagnosis result.  Run
pkgchk -n SUNWfmd to ensure that fault management software is installed
properly.
[server_cpu0] v:0x0000000000140750 p:0x001c1140750  cmp %l1, 0
Setting new inspection cpu: server_cpu0
simics> ptime
processor            steps         cycles  time [s]
server_cpu0  1168699608000  1168699608000   584.350

-----------------------------------------------------------------------------------------------------------------------------


2014-07-07 0:12 GMT+09:00 Cansu Kaynak <[email protected]>:

 Can you tell me after how many cycles you see this error when you start
> Simics from phase_000?
> You can check the cycles with the “ptime” command in SImics.
>
>   --
> Cansu
>
>  On 06 Jul 2014, at 16:52, J Ahn <[email protected]> wrote:
>
>  From the provided CloudSuite images, I am creating some additional
> phases of CloudStone application.
> The simulated system produces the ZFS messages as follows. (My host
> machine does not use Solaris with ZFS system.)
>
>
> ------------------------------------------------------------------------------------------------------------------------------------------------
> $] /opt/virtutech/simics-3.0.31/bin/simics -fast
> /home/flexus/workloads/cloudstone/2cpu/baseline/phase_000/simics/phase
> _000 -no-win
> Checking out a license... done: academic license.
>
>    +----------------+    Copyright 1998-2007 by Virtutech, All Rights
> Reserved
>   |   Virtutech    |    Version: Simics 3.0.31
>   |     Simics     |    Build: 1406  Host: amd64-linux
>   +----------------+
>     www.simics.com      "Virtutech" and "Simics" are trademarks of
> Virtutech AB
>
>  Use of this software is subject to appropriate license.
> Type 'copyright' for details on copyright.
> Type 'help help' for info on the on-line documentation.
>
>  simics> run-command-file create_phases.simics
> simics> c
>
>  SUNW-MSG-ID: FMD-8000-11, TYPE: Defect, VER: 1, SEVERITY: Minor
> EVENT-TIME: Sat May 19 03:13:22 GMT 2012
> PLATFORM: SUNW,Sun-Fire, CSN: -, HOSTNAME: besim
> SOURCE: zfs-diagnosis, REV: 1.0
> EVENT-ID: 6eac8d7b-7d99-6aaf-c063-cdbea30a0002
> DESC: A Solaris Fault Manager component generated a diagnosis for which no
> message summary exists.  Refer to http://sun.com/msg/FMD-8000-11 fo$ more
> information.
> AUTO-RESPONSE: The diagnosis has been saved in the fault log for
> examination by Sun.
> IMPACT: The fault log will need to be manually examined using fmdump(1M)
> in order to determine if any human response is required.
> REC-ACTION: Use fmdump -v -u <EVENT-ID> to view the diagnosis result.  Run
> pkgchk -n SUNWfmd to ensure that fault management software is insta$led
> properly.
> [server_hme0 info] We don't own next Rx descriptor
> [server_hme0 info] We don't own next Rx descriptor
> [server_hme0 info] We don't own next Rx descriptor
>
>
> ------------------------------------------------------------------------------------------------------------------------------------------------
>
>
> 2014-07-06 23:47 GMT+09:00 Cansu Kaynak <[email protected]>:
>
>> Dear Jeongseob,
>>
>>  Can you please copy and paste the error messages?
>> And please indicate if you see the messages in Simics or on your host
>> machine.
>>
>>  --
>> Cansu
>>
>>  On 06 Jul 2014, at 16:35, J Ahn <[email protected]> wrote:
>>
>>  I am also wondering whether the ZFS diagnosis message affect the
>> simulated systems or not.
>> When creating some additional phases from the released phase-0, I often
>> see the ZFS diagnosis messages in many applications(Cassandra, Streaming,
>> Nutch, and etc) of CloudSuite.
>>
>>  Does anyone who experience the similar problems ?
>>
>>
>> 2014-07-06 19:14 GMT+09:00 J Ahn <[email protected]>:
>>
>>> Hi all,
>>>
>>>  When creating the phases for Cassandra with 16cpus, the java
>>> exceptions occurs as follows.
>>>
>>>
>>> ------------------------------------------------------------------------------------------------------------------
>>>
>>>  simics> c
>>>                                                                   [39/85]
>>>  37 0 0 404440 3141776 0 67  0  0  0  0  0  0  0 66 90 112043 253435
>>> 104683 91 8 1
>>>  78 0 0 404352 3129384 0 37  0  0  0  0  0  0  0 66 68 112976 256015
>>> 105540 91 8 1
>>>
>>>  SUNW-MSG-ID: FMD-8000-11, TYPE: Defect, VER: 1, SEVERITY: Minor
>>> EVENT-TIME: Thu Mar  1 03:40:13 GMT 2012
>>> PLATFORM: SUNW,Sun-Fire, CSN: -, HOSTNAME: server
>>> SOURCE: zfs-diagnosis, REV: 1.0
>>> EVENT-ID: e72c81aa-5c28-6b5f-9a35-886022834a70
>>> DESC: A Solaris Fault Manager component generated a diagnosis for which
>>> no message summary exists.  Refer to http://sun.com/msg/FMD-8000-11 fo$
>>>  more information.
>>> AUTO-RESPONSE: The diagnosis has been saved in the fault log for
>>> examination by Sun.
>>> IMPACT: The fault log will need to be manually examined using fmdump(1M)
>>> in order to determine if any human response is required.
>>> REC-ACTION: Use fmdump -v -u <EVENT-ID> to view the diagnosis result.
>>>  Run pkgchk -n SUNWfmd to ensure that fault management software is insta$
>>> led properly.
>>> log4j:ERROR Failed to flush writer,
>>> java.io.IOException: Bad exchange descriptor
>>>         at java.io.FileOutputStream.writeBytes(Native Method)
>>>         at java.io.FileOutputStream.write(FileOutputStream.java:260)
>>>         at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:202)
>>>         at
>>> sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:272)
>>>         at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:276)
>>>         at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:122)
>>>         at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:212)
>>>         at
>>> org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59)
>>>         at
>>> org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324)
>>>         at
>>> org.apache.log4j.RollingFileAppender.subAppend(RollingFileAppender.java:276)
>>>         at
>>> org.apache.log4j.WriterAppender.append(WriterAppender.java:162)
>>>         at
>>> org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251)
>>>         at
>>> org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66)
>>>         at org.apache.log4j.Category.callAppenders(Category.java:206)
>>>         at org.apache.log4j.Category.forcedLog(Category.java:391)
>>>         at org.apache.log4j.Category.log(Category.java:856)
>>>         at org.slf4j.impl.Log4jLoggerAdapter.info
>>> <http://org.slf4j.impl.log4jloggeradapter.info/>
>>> (Log4jLoggerAdapter.java:305)
>>>         at
>>> org.apache.cassandra.db.Memtable.writeSortedContents(Memtable.java:164)
>>>         at org.apache.cassandra.db.Memtable.access$000(Memtable.java:51)
>>>         at
>>> org.apache.cassandra.db.Memtable$1.runMayThrow(Memtable.java:176)
>>>         at
>>> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:30)
>>>         at
>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>>>         at
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>>>         at java.lang.Thread.run(Thread.java:619)
>>>  56 0 0 400560 3104880 28 154 0 0  0  0  0 59  0 89 53 108956 236869
>>> 107265 88 10 1
>>>  [server_sn0 info] Dropping packet to 127.0.0.16 - no route.
>>>  64 0 0 400048 3083632 35 143 0 0  0  0  0 79  0 74 54 115209 245856
>>> 109447 91 7 1
>>>  120 sec: 1354816 operations; 12048.5 current ops/sec; [UPDATE
>>> AverageLatency(ms)=1.71] [READ AverageLatency(ms)=23.3]
>>>  41 0 0 407048 3073368 93 575 0 0  0  0  0 184 0 48 45 111381 247377
>>> 105177 91 8 1
>>> Feb 29 22:40:16 server svc.startd[9]: system/fmd:default failed:
>>> transitioned to maintenance (see 'svcs -xv' for details)
>>>  33 0 0 406352 3035792 40 348 0 0  0  0  0 59  0 71 65 108814 240641
>>> 100231 91 7 2
>>>  0 0 0 406344 3035784 0  23  0  0  0  0  0  0  0 63 55 114383 246439
>>> 104711 91 8 1
>>>  67 0 0 417304 3022168 0 55  0  0  0  0  0  0  0 62 91 110965 237690
>>> 105908 91 8 2
>>>  60 0 0 416184 3000560 0 43  0  0  0  0  0  2  0 65 65 112764 253245
>>> 107545 90 8 2
>>>  46 0 0 414992 2896872 0 163 0  0  0  0  0  0  0 83 68 109318 272641
>>> 104447 90 8 2
>>>
>>>
>>> ------------------------------------------------------------------------------------------------------------------
>>>
>>>  I did not modify the images provided with EPFL at all. Could you
>>> experience the similar problem?
>>>
>>>  Best regards,
>>> Jeongseob
>>>
>>
>>
>>
>
>

Reply via email to