[jira] [Updated] (IGNITE-11729) Low description for lost policy functional

2019-04-11 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11729?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11729:

Description: 
Current description in 
https://apacheignite.readme.io/docs/partition-loss-policies seems not covered 
persistence case and text description is not clear

Probably we need to transform it into two tables (in-memory, persistence)
Rows - cache_write, cache_read, cache_remove,sql_read, sql_write
Columns - READ_ONLY_SAFE ,READ_ONLY_ALL, READ_WRITE_SAFE, etc

{code:java}
Policies
Ignite supports the following PartitionLossPolicies:

READ_ONLY_SAFE - all writes to a cache/table will fail with an exception. Reads 
will only be allowed for entries belonging to survived/alive partitions. Reads 
from lost partitions will fail with an exception.
READ_ONLY_ALL - reads are allowed from any partition including the lost ones. 
An exception is thrown in an attempt to write to any partition. The result of 
reading from a lost partition is undefined and may be different on different 
nodes in the cluster.
READ_WRITE_SAFE - all reads and writes are allowed for entries in 
survived/alive partitions. All reads and writes of entries belonging to the 
lost partitions will fail with an exception.
READ_WRITE_ALL - all reads and writes will proceed as if all partitions were in 
a consistent state (as if no partition loss happened). The result of reading 
from a lost partition is undefined and may be different on different nodes in 
the cluster.
IGNORE - this mode never marks a lost partition as lost, pretending that no 
partition loss has happened and clearing the partition loss state right away. 
Technically, the partition will not be added to the collection of 
lostPartitions which is the main difference from READ_WRITE_ALL mode. IGNORE 
mode is used by default.
{code}


  was:
Current description in 
https://apacheignite.readme.io/docs/partition-loss-policies seems not covered 
persistence case and text description is not clean

Probably we need to transform it into two tables (in-memory, persistence)
Rows - cache_write, cache_read, cache_remove,sql_read, sql_write
Columns - READ_ONLY_SAFE ,READ_ONLY_ALL, READ_WRITE_SAFE, etc

{code:java}
Policies
Ignite supports the following PartitionLossPolicies:

READ_ONLY_SAFE - all writes to a cache/table will fail with an exception. Reads 
will only be allowed for entries belonging to survived/alive partitions. Reads 
from lost partitions will fail with an exception.
READ_ONLY_ALL - reads are allowed from any partition including the lost ones. 
An exception is thrown in an attempt to write to any partition. The result of 
reading from a lost partition is undefined and may be different on different 
nodes in the cluster.
READ_WRITE_SAFE - all reads and writes are allowed for entries in 
survived/alive partitions. All reads and writes of entries belonging to the 
lost partitions will fail with an exception.
READ_WRITE_ALL - all reads and writes will proceed as if all partitions were in 
a consistent state (as if no partition loss happened). The result of reading 
from a lost partition is undefined and may be different on different nodes in 
the cluster.
IGNORE - this mode never marks a lost partition as lost, pretending that no 
partition loss has happened and clearing the partition loss state right away. 
Technically, the partition will not be added to the collection of 
lostPartitions which is the main difference from READ_WRITE_ALL mode. IGNORE 
mode is used by default.
{code}



> Low description for lost policy functional
> --
>
> Key: IGNITE-11729
> URL: https://issues.apache.org/jira/browse/IGNITE-11729
> Project: Ignite
>  Issue Type: Improvement
>  Components: documentation
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Major
> Fix For: 3.0, 2.8
>
>
> Current description in 
> https://apacheignite.readme.io/docs/partition-loss-policies seems not covered 
> persistence case and text description is not clear
> Probably we need to transform it into two tables (in-memory, persistence)
> Rows - cache_write, cache_read, cache_remove,sql_read, sql_write
> Columns - READ_ONLY_SAFE ,READ_ONLY_ALL, READ_WRITE_SAFE, etc
> {code:java}
> Policies
> Ignite supports the following PartitionLossPolicies:
> READ_ONLY_SAFE - all writes to a cache/table will fail with an exception. 
> Reads will only be allowed for entries belonging to survived/alive 
> partitions. Reads from lost partitions will fail with an exception.
> READ_ONLY_ALL - reads are allowed from any partition including the lost ones. 
> An exception is thrown in an attempt to write to any partition. The result of 
> reading from a lost partition is undefined and may be different on different 
> nodes in the cluster.
> READ_WRITE_SAFE - all reads and writes are allowed for 

[jira] [Created] (IGNITE-11729) Low description for lost policy functional

2019-04-11 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11729:
---

 Summary: Low description for lost policy functional
 Key: IGNITE-11729
 URL: https://issues.apache.org/jira/browse/IGNITE-11729
 Project: Ignite
  Issue Type: Improvement
  Components: documentation
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 3.0, 2.8


Current description in 
https://apacheignite.readme.io/docs/partition-loss-policies seems not covered 
persistence case and text description is not clean

Probably we need to transform it into two tables (in-memory, persistence)
Rows - cache_write, cache_read, cache_remove,sql_read, sql_write
Columns - READ_ONLY_SAFE ,READ_ONLY_ALL, READ_WRITE_SAFE, etc

{code:java}
Policies
Ignite supports the following PartitionLossPolicies:

READ_ONLY_SAFE - all writes to a cache/table will fail with an exception. Reads 
will only be allowed for entries belonging to survived/alive partitions. Reads 
from lost partitions will fail with an exception.
READ_ONLY_ALL - reads are allowed from any partition including the lost ones. 
An exception is thrown in an attempt to write to any partition. The result of 
reading from a lost partition is undefined and may be different on different 
nodes in the cluster.
READ_WRITE_SAFE - all reads and writes are allowed for entries in 
survived/alive partitions. All reads and writes of entries belonging to the 
lost partitions will fail with an exception.
READ_WRITE_ALL - all reads and writes will proceed as if all partitions were in 
a consistent state (as if no partition loss happened). The result of reading 
from a lost partition is undefined and may be different on different nodes in 
the cluster.
IGNORE - this mode never marks a lost partition as lost, pretending that no 
partition loss has happened and clearing the partition loss state right away. 
Technically, the partition will not be added to the collection of 
lostPartitions which is the main difference from READ_WRITE_ALL mode. IGNORE 
mode is used by default.
{code}




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-11322) [USABILITY] Extend Node FAILED message by add consistentId if it exist

2019-02-14 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11322:
---

 Summary: [USABILITY] Extend Node FAILED message by add 
consistentId if it exist
 Key: IGNITE-11322
 URL: https://issues.apache.org/jira/browse/IGNITE-11322
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8


Now I having only this 
[GridDiscoveryManager] Node FAILED: TcpDiscoveryNode 
[id=f8cd73a1-8da5-4a07-b298-55634dd7c9f8, addrs=ArrayList [127.0.0.1], 
sockAddrs=HashSet [/127.0.0.1:47500], discPort=47500, order=1, intOrder=1, 
lastExchangeTime=1550141566893, loc=false, isClient=false]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-11302) idleConnectionTimeout TcpComm different on server and client (client default > server custom) lead to wait until client timeout on server side

2019-02-12 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11302:
---

 Summary: idleConnectionTimeout TcpComm different on server and 
client (client default > server custom) lead to wait until client timeout on 
server side
 Key: IGNITE-11302
 URL: https://issues.apache.org/jira/browse/IGNITE-11302
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.5
Reporter: ARomantsov
 Fix For: 2.8


Server config:





Client config





Server wait until default idleConnectionTimeout (10 m) for client fail.
If both config with idleConnectionTimeout=1 s - ignite worked according to 
config




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (IGNITE-11263) If host in TcpDiscoveryVmIpFinder is unreachable - server hang on start

2019-02-08 Thread ARomantsov (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-11263?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16763524#comment-16763524
 ] 

ARomantsov commented on IGNITE-11263:
-

Finally it start

[14:39:45,252][INFO][main][TcpDiscoverySpi] Successfully bound to TCP port 
[port=47500, localHost=0.0.0.0/0.0.0.0, 
locNodeId=6621cfd7-26a8-4bc0-a80a-88c6281aa118]
[14:50:45,910][WARNING][snapshot-scheduler-restats-#67][SnapshotScheduleProcessor]
 Snapshot schedule processor awaits for cluster activation.

> If host in TcpDiscoveryVmIpFinder is unreachable - server hang on start
> ---
>
> Key: IGNITE-11263
> URL: https://issues.apache.org/jira/browse/IGNITE-11263
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> If one of servers in  discovery ipFinder is unreachable - server hang on start
> repeat locally with commands
> sudo iptables -A INPUT DROP
> sudo iptables -A OUTPUT DROP
> {code:java}
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
>  class="org.apache.ignite.spi.discovery.tcp.ipfinder.vm.TcpDiscoveryVmIpFinder">
> 
> 
> {code}
> Server logs end with
> {code:java}
> [14:39:44,668][INFO][main][GridCacheDatabaseSharedManager] Read checkpoint 
> status [startMarker=null, endMarker=null]
> [14:39:44,669][INFO][main][GridCacheDatabaseSharedManager] Applying lost 
> cache updates since last checkpoint record [lastMarked=FileWALPointer [idx=0, 
> fileOff=0, len=0], lastCheckpointId=----]
> [14:39:44,700][INFO][main][GridCacheDatabaseSharedManager] Finished applying 
> WAL changes [updatesApplied=0, time=30 ms]
> [14:39:44,700][INFO][main][GridCacheDatabaseSharedManager] Restoring 
> partition state for local groups.
> [14:39:44,705][INFO][main][GridCacheDatabaseSharedManager] Finished restoring 
> partition state for local groups [groupsProcessed11partitionsProcessed=0, 
> time=10ms]
> [14:39:45,252][INFO][main][TcpDiscoverySpi] Successfully bound to TCP port 
> [port=47500, localHost=0.0.0.0/0.0.0.0, 
> locNodeId=6621cfd7-26a8-4bc0-a80a-88c6281aa118]
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (IGNITE-11263) If host in TcpDiscoveryVmIpFinder is unreachable - server hang on start

2019-02-08 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11263?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov resolved IGNITE-11263.
-
Resolution: Invalid

> If host in TcpDiscoveryVmIpFinder is unreachable - server hang on start
> ---
>
> Key: IGNITE-11263
> URL: https://issues.apache.org/jira/browse/IGNITE-11263
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> If one of servers in  discovery ipFinder is unreachable - server hang on start
> repeat locally with commands
> sudo iptables -A INPUT DROP
> sudo iptables -A OUTPUT DROP
> {code:java}
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
>  class="org.apache.ignite.spi.discovery.tcp.ipfinder.vm.TcpDiscoveryVmIpFinder">
> 
> 
> {code}
> Server logs end with
> {code:java}
> [14:39:44,668][INFO][main][GridCacheDatabaseSharedManager] Read checkpoint 
> status [startMarker=null, endMarker=null]
> [14:39:44,669][INFO][main][GridCacheDatabaseSharedManager] Applying lost 
> cache updates since last checkpoint record [lastMarked=FileWALPointer [idx=0, 
> fileOff=0, len=0], lastCheckpointId=----]
> [14:39:44,700][INFO][main][GridCacheDatabaseSharedManager] Finished applying 
> WAL changes [updatesApplied=0, time=30 ms]
> [14:39:44,700][INFO][main][GridCacheDatabaseSharedManager] Restoring 
> partition state for local groups.
> [14:39:44,705][INFO][main][GridCacheDatabaseSharedManager] Finished restoring 
> partition state for local groups [groupsProcessed11partitionsProcessed=0, 
> time=10ms]
> [14:39:45,252][INFO][main][TcpDiscoverySpi] Successfully bound to TCP port 
> [port=47500, localHost=0.0.0.0/0.0.0.0, 
> locNodeId=6621cfd7-26a8-4bc0-a80a-88c6281aa118]
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-11263) If host in TcpDiscoveryVmIpFinder is unreachable - server hang on start

2019-02-08 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11263:
---

 Summary: If host in TcpDiscoveryVmIpFinder is unreachable - server 
hang on start
 Key: IGNITE-11263
 URL: https://issues.apache.org/jira/browse/IGNITE-11263
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.5
Reporter: ARomantsov
 Fix For: 2.8


If one of servers in  discovery ipFinder is unreachable - server hang on start

repeat locally with commands
sudo iptables -A INPUT DROP
sudo iptables -A OUTPUT DROP

{code:java}














{code}


Server logs end with


{code:java}

[14:39:44,668][INFO][main][GridCacheDatabaseSharedManager] Read checkpoint 
status [startMarker=null, endMarker=null]
[14:39:44,669][INFO][main][GridCacheDatabaseSharedManager] Applying lost cache 
updates since last checkpoint record [lastMarked=FileWALPointer [idx=0, 
fileOff=0, len=0], lastCheckpointId=----]
[14:39:44,700][INFO][main][GridCacheDatabaseSharedManager] Finished applying 
WAL changes [updatesApplied=0, time=30 ms]
[14:39:44,700][INFO][main][GridCacheDatabaseSharedManager] Restoring partition 
state for local groups.
[14:39:44,705][INFO][main][GridCacheDatabaseSharedManager] Finished restoring 
partition state for local groups [groupsProcessed11partitionsProcessed=0, 
time=10ms]
[14:39:45,252][INFO][main][TcpDiscoverySpi] Successfully bound to TCP port 
[port=47500, localHost=0.0.0.0/0.0.0.0, 
locNodeId=6621cfd7-26a8-4bc0-a80a-88c6281aa118]

{code}






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (IGNITE-11153) Support Java 11 for control.sh & ignitevisorcmd.sh: remove odd --add-modules for java.xml.bind

2019-02-07 Thread ARomantsov (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-11153?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16762589#comment-16762589
 ] 

ARomantsov commented on IGNITE-11153:
-

Hello, thank for adding support for Java11 in utilities. I've checked changes 
code in bat scripts and have builded ignite from linked PR and it works correct

> Support Java 11 for control.sh & ignitevisorcmd.sh: remove odd --add-modules 
> for java.xml.bind
> --
>
> Key: IGNITE-11153
> URL: https://issues.apache.org/jira/browse/IGNITE-11153
> Project: Ignite
>  Issue Type: Task
>Affects Versions: 2.7
>Reporter: Dmitriy Pavlov
>Assignee: Peter Ivanov
>Priority: Major
> Fix For: 2.8
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> JAXB module was removed from Java 11.
> Please adopt ignite control.sh in the same way as it is done for ignite.sh/bat
> {noformat}
> elif [ $version -gt 8 ] && [ $version -lt 11 ]; then
> JVM_OPTS="\
> -XX:+AggressiveOpts \
> --add-exports=java.base/jdk.internal.misc=ALL-UNNAMED \
> --add-exports=java.base/sun.nio.ch=ALL-UNNAMED \
> --add-exports=java.management/com.sun.jmx.mbeanserver=ALL-UNNAMED \
> --add-exports=jdk.internal.jvmstat/sun.jvmstat.monitor=ALL-UNNAMED \
> 
> --add-exports=java.base/sun.reflect.generics.reflectiveObjects=ALL-UNNAMED \
> --illegal-access=permit \
> --add-modules=java.transaction \
> --add-modules=java.xml.bind \
> ${JVM_OPTS}"
> elif [ $version -eq 11 ] ; then
> JVM_OPTS="\
> --add-exports=java.base/jdk.internal.misc=ALL-UNNAMED \
> --add-exports=java.base/sun.nio.ch=ALL-UNNAMED \
> --add-exports=java.management/com.sun.jmx.mbeanserver=ALL-UNNAMED \
> --add-exports=jdk.internal.jvmstat/sun.jvmstat.monitor=ALL-UNNAMED \
> 
> --add-exports=java.base/sun.reflect.generics.reflectiveObjects=ALL-UNNAMED \
> --illegal-access=permit \
> ${JVM_OPTS}"
> {noformat}
> But control.sh for master still declares   --add-modules java.xml.bind \
> {noformat}
> JVM_OPTS="--add-exports java.base/jdk.internal.misc=ALL-UNNAMED \
>   --add-exports java.base/sun.nio.ch=ALL-UNNAMED \
>   --add-exports java.management/com.sun.jmx.mbeanserver=ALL-UNNAMED \
>   --add-exports jdk.internal.jvmstat/sun.jvmstat.monitor=ALL-UNNAMED \
>   --add-modules java.xml.bind \
>   ${JVM_OPTS}"
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-11243) Not working control.sh / control.bat in master NPE in output

2019-02-07 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11243:
---

 Summary: Not working control.sh / control.bat in master NPE in 
output
 Key: IGNITE-11243
 URL: https://issues.apache.org/jira/browse/IGNITE-11243
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.8
Reporter: ARomantsov
 Fix For: 2.8


./bin/control.sh  --host --port --baseline
Cluster state: active
Error: java.lang.NullPointerException

control.bat --host  --port  --baseline
Cluster state: active
Error: java.lang.NullPointerException
Press any key to continue . . .



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-11243) Not working control.sh / control.bat in master NPE in output

2019-02-07 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11243?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11243:

Description: 
./bin/control.sh  --host --port --baseline
Cluster state: active
Error: java.lang.NullPointerException

control.bat --host  --port  --baseline
Cluster state: active
Error: java.lang.NullPointerException
Press any key to continue . . .

No info in cluster logs matched with call, look like problem in utility run

  was:
./bin/control.sh  --host --port --baseline
Cluster state: active
Error: java.lang.NullPointerException

control.bat --host  --port  --baseline
Cluster state: active
Error: java.lang.NullPointerException
Press any key to continue . . .


> Not working control.sh / control.bat in master NPE in output
> 
>
> Key: IGNITE-11243
> URL: https://issues.apache.org/jira/browse/IGNITE-11243
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Blocker
> Fix For: 2.8
>
>
> ./bin/control.sh  --host --port --baseline
> Cluster state: active
> Error: java.lang.NullPointerException
> control.bat --host  --port  --baseline
> Cluster state: active
> Error: java.lang.NullPointerException
> Press any key to continue . . .
> No info in cluster logs matched with call, look like problem in utility run



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-11217) After cluster kill and restart with new jvm option - one of node fail with failure type - Failed to read WAL record at position

2019-02-05 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11217:
---

 Summary: After cluster kill and restart with new jvm option - one 
of node fail with failure type - Failed to read WAL record at position 
 Key: IGNITE-11217
 URL: https://issues.apache.org/jira/browse/IGNITE-11217
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8


Test scenario:

- Start two server load
- execute operations (put/put-all/streamer/remove)
- execute client that collect checksum
- kill nodes
- add JVM options to server - DIGNITE_WAL_LOG_TX_RECORDS=true, 
DIGNITE_RECOVERY_VERBOSE_LOGGING
- start nodes
- activate them
Found that one node fail


{code:java}
[23:16:06,230][INFO][main][GridCacheProcessor] Finished restoring partition 
state for local groups [groupsProcessed=45, partitionsProcessed=764, time=557ms]
[23:16:06,280][INFO][main][FileWriteAheadLogManager] Stopping WAL iteration due 
to an exception: Failed to read WAL record at position: 27700852 size: 
67108864, ptr=FileWALPointer [idx=0, fileOff=27700852, len=0]
[23:16:06,298][SEVERE][main][IgniteKernal] Exception during start processors, 
node will be stopped and close connections
class org.apache.ignite.IgniteCheckedException: Failed to read WAL record at 
position: 27700852 size: 67108864
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.handleRecordException(AbstractWalRecordsIterator.java:305)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$RecordsIterator.handleRecordException(FileWriteAheadLogManager.java:2691)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:269)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advance(AbstractWalRecordsIterator.java:158)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:127)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:53)
at 
org.apache.ignite.internal.util.GridCloseableIteratorAdapter.nextX(GridCloseableIteratorAdapter.java:41)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager.tailPointer(GridCacheDatabaseSharedManager.java:1984)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager.startMemoryRestore(GridCacheDatabaseSharedManager.java:1958)
at org.apache.ignite.internal.IgniteKernal.start(IgniteKernal.java:1050)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance.start0(IgnitionEx.java:2040)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance.start(IgnitionEx.java:1732)
at org.apache.ignite.internal.IgnitionEx.start0(IgnitionEx.java:1158)
at 
org.apache.ignite.internal.IgnitionEx.startConfigurations(IgnitionEx.java:1076)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:962)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:861)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:731)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:700)
at org.apache.ignite.Ignition.start(Ignition.java:348)
at 
org.apache.ignite.startup.cmdline.CommandLineStartup.main(CommandLineStartup.java:301)
Caused by: class org.apache.ignite.IgniteCheckedException: Failed to read WAL 
record at position: 27700852 size: 67108864
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:395)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer.readRecord(RecordV2Serializer.java:236)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:254)
... 17 more
Caused by: java.io.IOException: class org.apache.ignite.IgniteCheckedException: 
Unsupported protocol version: 34
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordDataV1Serializer.readVersion(RecordDataV1Serializer.java:2010)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordDataV2Serializer.readMvccDataEntry(RecordDataV2Serializer.java:328)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordDataV2Serializer.readPlainRecord(RecordDataV2Serializer.java:162)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordDataV1Serializer.readRecord(RecordDataV1Serializer.java:208)
at 

[jira] [Created] (IGNITE-11184) add example of ssl rest protocol on ignite

2019-02-04 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11184:
---

 Summary: add example of ssl rest protocol on ignite
 Key: IGNITE-11184
 URL: https://issues.apache.org/jira/browse/IGNITE-11184
 Project: Ignite
  Issue Type: Bug
  Components: documentation
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8


Add information about ssl Jetty configuration to ignite documentation about 
restApi 
https://apacheignite.readme.io/docs/rest-api#sample-jetty-xml-configuration



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-11179) One of two nodes fail by handler with message - threadName=ttl-cleanup-worker, blockedFor=16s

2019-02-03 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11179:
---

 Summary: One of two nodes fail by handler with message - 
threadName=ttl-cleanup-worker, blockedFor=16s
 Key: IGNITE-11179
 URL: https://issues.apache.org/jira/browse/IGNITE-11179
 Project: Ignite
  Issue Type: Bug
Reporter: ARomantsov


Start two nodes, one of them drop after Start caches in recovery mode



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-11179) One of two nodes fail by handler with message - threadName=ttl-cleanup-worker, blockedFor=16s

2019-02-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11179:

Fix Version/s: 2.8

> One of two nodes fail by handler with message - 
> threadName=ttl-cleanup-worker, blockedFor=16s
> -
>
> Key: IGNITE-11179
> URL: https://issues.apache.org/jira/browse/IGNITE-11179
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> Start two nodes, one of them drop after Start caches in recovery mode



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-11179) One of two nodes fail by handler with message - threadName=ttl-cleanup-worker, blockedFor=16s

2019-02-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11179:

Description: 
Start two nodes, one of them fail after block "start caches  in recovery mode"


{code:java}
[12:47:57,333][SEVERE][tcp-disco-msg-worker-#2][G] Blocked system-critical 
thread has been detected. This can lead to cluster-wide undefined behaviour 
[threadName=ttl-cleanup-worker, blockedFor=16s]
[12:47:57,334][WARNING][tcp-disco-msg-worker-#2][G] Thread 
[name="ttl-cleanup-worker-#45", id=63, state=WAITING, blockCnt=0, waitCnt=1]

[12:47:57,336][SEVERE][tcp-disco-msg-worker-#2][] Critical system error 
detected. Will be handled accordingly to configured handler 
[hnd=StopNodeOrHaltFailureHandler [tryStop=false, timeout=0, 
super=AbstractFailureHandler [ignoredFailureTypes=UnmodifiableSet 
[SYSTEM_WORKER_BLOCKED, SYSTEM_CRITICAL_OPERATION_TIMEOUT]]], 
failureCtx=FailureContext [type=SYSTEM_WORKER_BLOCKED, err=class 
o.a.i.IgniteException: GridWorker [name=ttl-cleanup-worker, 
igniteInstanceName=null, finished=false, heartbeatTs=1549187260799]]]
class org.apache.ignite.IgniteException: GridWorker [name=ttl-cleanup-worker, 
igniteInstanceName=null, finished=false, heartbeatTs=1549187260799]
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance$2.apply(IgnitionEx.java:1833)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance$2.apply(IgnitionEx.java:1828)
at 
org.apache.ignite.internal.worker.WorkersRegistry.onIdle(WorkersRegistry.java:233)
at 
org.apache.ignite.internal.util.worker.GridWorker.onIdle(GridWorker.java:297)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.lambda$new$0(ServerImpl.java:2680)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorker.body(ServerImpl.java:7218)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2717)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:120)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerThread.body(ServerImpl.java:7156)
at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
{code}


  was:
Start two nodes, one of them drop after Start caches in recovery mode


{code:java}
[12:47:57,333][SEVERE][tcp-disco-msg-worker-#2][G] Blocked system-critical 
thread has been detected. This can lead to cluster-wide undefined behaviour 
[threadName=ttl-cleanup-worker, blockedFor=16s]
[12:47:57,334][WARNING][tcp-disco-msg-worker-#2][G] Thread 
[name="ttl-cleanup-worker-#45", id=63, state=WAITING, blockCnt=0, waitCnt=1]

[12:47:57,336][SEVERE][tcp-disco-msg-worker-#2][] Critical system error 
detected. Will be handled accordingly to configured handler 
[hnd=StopNodeOrHaltFailureHandler [tryStop=false, timeout=0, 
super=AbstractFailureHandler [ignoredFailureTypes=UnmodifiableSet 
[SYSTEM_WORKER_BLOCKED, SYSTEM_CRITICAL_OPERATION_TIMEOUT]]], 
failureCtx=FailureContext [type=SYSTEM_WORKER_BLOCKED, err=class 
o.a.i.IgniteException: GridWorker [name=ttl-cleanup-worker, 
igniteInstanceName=null, finished=false, heartbeatTs=1549187260799]]]
class org.apache.ignite.IgniteException: GridWorker [name=ttl-cleanup-worker, 
igniteInstanceName=null, finished=false, heartbeatTs=1549187260799]
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance$2.apply(IgnitionEx.java:1833)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance$2.apply(IgnitionEx.java:1828)
at 
org.apache.ignite.internal.worker.WorkersRegistry.onIdle(WorkersRegistry.java:233)
at 
org.apache.ignite.internal.util.worker.GridWorker.onIdle(GridWorker.java:297)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.lambda$new$0(ServerImpl.java:2680)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorker.body(ServerImpl.java:7218)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2717)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:120)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerThread.body(ServerImpl.java:7156)
at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
{code}



> One of two nodes fail by handler with message - 
> threadName=ttl-cleanup-worker, blockedFor=16s
> -
>
> Key: IGNITE-11179
> URL: https://issues.apache.org/jira/browse/IGNITE-11179
> Project: Ignite
>  Issue Type: Bug
>  Components: data structures
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> Start two nodes, one of them fail after block "start caches  in recovery mode"
> 

[jira] [Updated] (IGNITE-11179) One of two nodes fail by handler with message - threadName=ttl-cleanup-worker, blockedFor=16s

2019-02-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11179:

Description: 
Start two nodes, one of them drop after Start caches in recovery mode


{code:java}
[12:47:57,333][SEVERE][tcp-disco-msg-worker-#2][G] Blocked system-critical 
thread has been detected. This can lead to cluster-wide undefined behaviour 
[threadName=ttl-cleanup-worker, blockedFor=16s]
[12:47:57,334][WARNING][tcp-disco-msg-worker-#2][G] Thread 
[name="ttl-cleanup-worker-#45", id=63, state=WAITING, blockCnt=0, waitCnt=1]

[12:47:57,336][SEVERE][tcp-disco-msg-worker-#2][] Critical system error 
detected. Will be handled accordingly to configured handler 
[hnd=StopNodeOrHaltFailureHandler [tryStop=false, timeout=0, 
super=AbstractFailureHandler [ignoredFailureTypes=UnmodifiableSet 
[SYSTEM_WORKER_BLOCKED, SYSTEM_CRITICAL_OPERATION_TIMEOUT]]], 
failureCtx=FailureContext [type=SYSTEM_WORKER_BLOCKED, err=class 
o.a.i.IgniteException: GridWorker [name=ttl-cleanup-worker, 
igniteInstanceName=null, finished=false, heartbeatTs=1549187260799]]]
class org.apache.ignite.IgniteException: GridWorker [name=ttl-cleanup-worker, 
igniteInstanceName=null, finished=false, heartbeatTs=1549187260799]
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance$2.apply(IgnitionEx.java:1833)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance$2.apply(IgnitionEx.java:1828)
at 
org.apache.ignite.internal.worker.WorkersRegistry.onIdle(WorkersRegistry.java:233)
at 
org.apache.ignite.internal.util.worker.GridWorker.onIdle(GridWorker.java:297)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.lambda$new$0(ServerImpl.java:2680)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorker.body(ServerImpl.java:7218)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2717)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:120)
at 
org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerThread.body(ServerImpl.java:7156)
at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
{code}


  was:Start two nodes, one of them drop after Start caches in recovery mode


> One of two nodes fail by handler with message - 
> threadName=ttl-cleanup-worker, blockedFor=16s
> -
>
> Key: IGNITE-11179
> URL: https://issues.apache.org/jira/browse/IGNITE-11179
> Project: Ignite
>  Issue Type: Bug
>  Components: data structures
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> Start two nodes, one of them drop after Start caches in recovery mode
> {code:java}
> [12:47:57,333][SEVERE][tcp-disco-msg-worker-#2][G] Blocked system-critical 
> thread has been detected. This can lead to cluster-wide undefined behaviour 
> [threadName=ttl-cleanup-worker, blockedFor=16s]
> [12:47:57,334][WARNING][tcp-disco-msg-worker-#2][G] Thread 
> [name="ttl-cleanup-worker-#45", id=63, state=WAITING, blockCnt=0, waitCnt=1]
> [12:47:57,336][SEVERE][tcp-disco-msg-worker-#2][] Critical system error 
> detected. Will be handled accordingly to configured handler 
> [hnd=StopNodeOrHaltFailureHandler [tryStop=false, timeout=0, 
> super=AbstractFailureHandler [ignoredFailureTypes=UnmodifiableSet 
> [SYSTEM_WORKER_BLOCKED, SYSTEM_CRITICAL_OPERATION_TIMEOUT]]], 
> failureCtx=FailureContext [type=SYSTEM_WORKER_BLOCKED, err=class 
> o.a.i.IgniteException: GridWorker [name=ttl-cleanup-worker, 
> igniteInstanceName=null, finished=false, heartbeatTs=1549187260799]]]
> class org.apache.ignite.IgniteException: GridWorker [name=ttl-cleanup-worker, 
> igniteInstanceName=null, finished=false, heartbeatTs=1549187260799]
>   at 
> org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance$2.apply(IgnitionEx.java:1833)
>   at 
> org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance$2.apply(IgnitionEx.java:1828)
>   at 
> org.apache.ignite.internal.worker.WorkersRegistry.onIdle(WorkersRegistry.java:233)
>   at 
> org.apache.ignite.internal.util.worker.GridWorker.onIdle(GridWorker.java:297)
>   at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.lambda$new$0(ServerImpl.java:2680)
>   at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorker.body(ServerImpl.java:7218)
>   at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2717)
>   at 
> org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:120)
>   at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerThread.body(ServerImpl.java:7156)
>   at 

[jira] [Updated] (IGNITE-11179) One of two nodes fail by handler with message - threadName=ttl-cleanup-worker, blockedFor=16s

2019-02-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11179:

Affects Version/s: 2.8

> One of two nodes fail by handler with message - 
> threadName=ttl-cleanup-worker, blockedFor=16s
> -
>
> Key: IGNITE-11179
> URL: https://issues.apache.org/jira/browse/IGNITE-11179
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Critical
>
> Start two nodes, one of them drop after Start caches in recovery mode



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-11179) One of two nodes fail by handler with message - threadName=ttl-cleanup-worker, blockedFor=16s

2019-02-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11179:

Component/s: data structures

> One of two nodes fail by handler with message - 
> threadName=ttl-cleanup-worker, blockedFor=16s
> -
>
> Key: IGNITE-11179
> URL: https://issues.apache.org/jira/browse/IGNITE-11179
> Project: Ignite
>  Issue Type: Bug
>  Components: data structures
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> Start two nodes, one of them drop after Start caches in recovery mode



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-11103) "Control utility --cache idle_verify --dump --cache-filter ALL" comand result doesn't contain ignite-sys-cache group

2019-01-28 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11103:

Fix Version/s: 2.8

> "Control utility --cache idle_verify --dump --cache-filter ALL" comand result 
> doesn't contain ignite-sys-cache group
> 
>
> Key: IGNITE-11103
> URL: https://issues.apache.org/jira/browse/IGNITE-11103
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> Look at functional add in https://issues.apache.org/jira/browse/IGNITE-9980 
> and find that issue



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-11103) "Control utility --cache idle_verify --dump --cache-filter ALL" comand result doesn't contain ignite-sys-cache group

2019-01-28 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-11103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-11103:

Affects Version/s: 2.8

> "Control utility --cache idle_verify --dump --cache-filter ALL" comand result 
> doesn't contain ignite-sys-cache group
> 
>
> Key: IGNITE-11103
> URL: https://issues.apache.org/jira/browse/IGNITE-11103
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Major
>
> Look at functional add in https://issues.apache.org/jira/browse/IGNITE-9980 
> and find that issue



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-11103) "Control utility --cache idle_verify --dump --cache-filter ALL" comand result doesn't contain ignite-sys-cache group

2019-01-28 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-11103:
---

 Summary: "Control utility --cache idle_verify --dump 
--cache-filter ALL" comand result doesn't contain ignite-sys-cache group
 Key: IGNITE-11103
 URL: https://issues.apache.org/jira/browse/IGNITE-11103
 Project: Ignite
  Issue Type: Bug
Reporter: ARomantsov


Look at functional add in https://issues.apache.org/jira/browse/IGNITE-9980 and 
find that issue



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10975) when ssl not correct configurate make cu error more undestandable

2019-01-18 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10975:

Affects Version/s: 2.5

> when ssl not correct configurate make cu error more undestandable
> -
>
> Key: IGNITE-10975
> URL: https://issues.apache.org/jira/browse/IGNITE-10975
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: ARomantsov
>Priority: Major
>
> Now CU return
> Connection to cluster failed.
> Error: Latest topology update failed 
> this error also appear when you try to connect on unexist cluster , at that 
> point gain it with bad ssl that add little bit confusion



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10975) when ssl not correct configurate make cu error more undestandable

2019-01-18 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10975:
---

 Summary: when ssl not correct configurate make cu error more 
undestandable
 Key: IGNITE-10975
 URL: https://issues.apache.org/jira/browse/IGNITE-10975
 Project: Ignite
  Issue Type: Bug
Reporter: ARomantsov


Now CU return
Connection to cluster failed.
Error: Latest topology update failed 

this error also appear when you try to connect on unexist cluster , at that 
point gain it with bad ssl that add little bit confusion



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10975) when ssl not correct configurate make cu error more undestandable

2019-01-18 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10975:

Fix Version/s: 2.8

> when ssl not correct configurate make cu error more undestandable
> -
>
> Key: IGNITE-10975
> URL: https://issues.apache.org/jira/browse/IGNITE-10975
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> Now CU return
> Connection to cluster failed.
> Error: Latest topology update failed 
> this error also appear when you try to connect on unexist cluster , at that 
> point gain it with bad ssl that add little bit confusion



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10938) After restart cluster with non-blt nodes - they left by handler

2019-01-15 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10938:
---

 Summary: After restart cluster with non-blt nodes - they left by 
handler
 Key: IGNITE-10938
 URL: https://issues.apache.org/jira/browse/IGNITE-10938
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.8
Reporter: ARomantsov
 Fix For: 2.8


I have cluster wherein topology contain blt and non-blt nodes, but after 
restart - nodes left by handler

java.lang.IllegalStateException: Unable to find consistentId by UUID




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10929) After huge load on cluster and restart with walCompactionEnabled=True warning on log

2019-01-14 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10929:
---

 Summary: After huge load on cluster and restart with 
walCompactionEnabled=True warning on log
 Key: IGNITE-10929
 URL: https://issues.apache.org/jira/browse/IGNITE-10929
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.5
Reporter: ARomantsov
 Fix For: 2.8



{code:java}
[15:08:14,610][WARNING][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0001.wal, exists: 
false
[15:08:15,661][WARNING][wal-file-compressor-%null%-0-#66][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0005.wal, exists: 
false
[15:08:16,540][WARNING][wal-file-compressor-%null%-2-#69][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0009.wal, exists: 
false
[15:08:17,354][WARNING][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0013.wal, exists: 
false
[15:08:18,161][WARNING][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0017.wal, exists: 
false
[15:08:18,161][WARNING][wal-file-compressor-%null%-2-#69][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0017.wal, exists: 
false
[15:08:18,161][WARNING][wal-file-compressor-%null%-2-#69][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0018.wal, exists: 
false
[15:08:18,987][WARNING][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0020.wal, exists: 
false
[15:08:18,987][WARNING][wal-file-compressor-%null%-2-#69][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0020.wal, exists: 
false
[15:08:18,998][WARNING][wal-file-compressor-%null%-2-#69][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0023.wal, exists: 
false
[15:08:23,211][WARNING][wal-file-compressor-%null%-2-#69][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0028.wal, exists: 
false
[15:08:23,211][WARNING][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0029.wal, exists: 
false
[15:08:24,264][WARNING][wal-file-compressor-%null%-0-#66][FileWriteAheadLogManager]
 Failed to remove obsolete WAL segment (make sure the process has enough 
rights):  my_path/work/db/wal/archive/node_1_1/0033.wal, exists: 
false
{code}




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10928) After huge load on cluster and restart with walCompactionEnabled=True errors on log

2019-01-14 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10928:

Affects Version/s: 2.5

> After huge load on cluster and restart with walCompactionEnabled=True errors 
> on log
> ---
>
> Key: IGNITE-10928
> URL: https://issues.apache.org/jira/browse/IGNITE-10928
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: ARomantsov
>Priority: Critical
>
> {code:java}
> 
>  class="org.apache.ignite.configuration.DataRegionConfiguration">
> 
> 
> 
> 
> 
> {code}
> {code:java}
> [15:30:56,809][INFO][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
>  Stopping WAL iteration due to an exception: Failed to read WAL record at 
> position: 28310114 size: -1, ptr=FileWALPointer [idx=35, fileOff=28310114, 
> len=0]
> [15:30:56,811][INFO][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
>  Stopping WAL iteration due to an exception: Failed to read WAL record at 
> position: 28303753 size: -1, ptr=FileWALPointer [idx=36, fileOff=28303753, 
> len=0]
> [15:30:56,811][SEVERE][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
>  Compression of WAL segment [idx=35] was skipped due to unexpected error
> class org.apache.ignite.IgniteCheckedException: Failed to read WAL record at 
> position: 28310114 size: -1
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.handleRecordException(AbstractWalRecordsIterator.java:292)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:258)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advance(AbstractWalRecordsIterator.java:154)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.SingleSegmentLogicalRecordsIterator.advance(SingleSegmentLogicalRecordsIterator.java:119)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:123)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:52)
> at 
> org.apache.ignite.internal.util.GridCloseableIteratorAdapter.nextX(GridCloseableIteratorAdapter.java:41)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.compressSegmentToFile(FileWriteAheadLogManager.java:2039)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body0(FileWriteAheadLogManager.java:1974)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body(FileWriteAheadLogManager.java:1950)
> at 
> org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: class org.apache.ignite.IgniteCheckedException: Failed to read WAL 
> record at position: 28310114 size: -1
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:394)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer.readRecord(RecordV2Serializer.java:235)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:243)
> ... 10 more
> Caused by: java.nio.channels.ClosedByInterruptException
> at 
> java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
> at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:164)
> at 
> org.apache.ignite.internal.processors.cache.persistence.file.RandomAccessFileIO.read(RandomAccessFileIO.java:58)
> at 
> org.apache.ignite.internal.processors.cache.persistence.file.FileIODecorator.read(FileIODecorator.java:51)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.SimpleFileInput.ensure(SimpleFileInput.java:119)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.ensure(FileInput.java:89)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.readFully(FileInput.java:152)
> at 
> 

[jira] [Updated] (IGNITE-10928) After huge load on cluster and restart with walCompactionEnabled=True errors on log

2019-01-14 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10928:

Component/s: data structures

> After huge load on cluster and restart with walCompactionEnabled=True errors 
> on log
> ---
>
> Key: IGNITE-10928
> URL: https://issues.apache.org/jira/browse/IGNITE-10928
> Project: Ignite
>  Issue Type: Bug
>  Components: data structures
>Affects Versions: 2.5
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> {code:java}
> 
>  class="org.apache.ignite.configuration.DataRegionConfiguration">
> 
> 
> 
> 
> 
> {code}
> {code:java}
> [15:30:56,809][INFO][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
>  Stopping WAL iteration due to an exception: Failed to read WAL record at 
> position: 28310114 size: -1, ptr=FileWALPointer [idx=35, fileOff=28310114, 
> len=0]
> [15:30:56,811][INFO][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
>  Stopping WAL iteration due to an exception: Failed to read WAL record at 
> position: 28303753 size: -1, ptr=FileWALPointer [idx=36, fileOff=28303753, 
> len=0]
> [15:30:56,811][SEVERE][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
>  Compression of WAL segment [idx=35] was skipped due to unexpected error
> class org.apache.ignite.IgniteCheckedException: Failed to read WAL record at 
> position: 28310114 size: -1
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.handleRecordException(AbstractWalRecordsIterator.java:292)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:258)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advance(AbstractWalRecordsIterator.java:154)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.SingleSegmentLogicalRecordsIterator.advance(SingleSegmentLogicalRecordsIterator.java:119)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:123)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:52)
> at 
> org.apache.ignite.internal.util.GridCloseableIteratorAdapter.nextX(GridCloseableIteratorAdapter.java:41)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.compressSegmentToFile(FileWriteAheadLogManager.java:2039)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body0(FileWriteAheadLogManager.java:1974)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body(FileWriteAheadLogManager.java:1950)
> at 
> org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: class org.apache.ignite.IgniteCheckedException: Failed to read WAL 
> record at position: 28310114 size: -1
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:394)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer.readRecord(RecordV2Serializer.java:235)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:243)
> ... 10 more
> Caused by: java.nio.channels.ClosedByInterruptException
> at 
> java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
> at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:164)
> at 
> org.apache.ignite.internal.processors.cache.persistence.file.RandomAccessFileIO.read(RandomAccessFileIO.java:58)
> at 
> org.apache.ignite.internal.processors.cache.persistence.file.FileIODecorator.read(FileIODecorator.java:51)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.SimpleFileInput.ensure(SimpleFileInput.java:119)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.ensure(FileInput.java:89)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.readFully(FileInput.java:152)
> at 
> 

[jira] [Updated] (IGNITE-10928) After huge load on cluster and restart with walCompactionEnabled=True errors on log

2019-01-14 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10928:

Fix Version/s: 2.8

> After huge load on cluster and restart with walCompactionEnabled=True errors 
> on log
> ---
>
> Key: IGNITE-10928
> URL: https://issues.apache.org/jira/browse/IGNITE-10928
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> {code:java}
> 
>  class="org.apache.ignite.configuration.DataRegionConfiguration">
> 
> 
> 
> 
> 
> {code}
> {code:java}
> [15:30:56,809][INFO][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
>  Stopping WAL iteration due to an exception: Failed to read WAL record at 
> position: 28310114 size: -1, ptr=FileWALPointer [idx=35, fileOff=28310114, 
> len=0]
> [15:30:56,811][INFO][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
>  Stopping WAL iteration due to an exception: Failed to read WAL record at 
> position: 28303753 size: -1, ptr=FileWALPointer [idx=36, fileOff=28303753, 
> len=0]
> [15:30:56,811][SEVERE][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
>  Compression of WAL segment [idx=35] was skipped due to unexpected error
> class org.apache.ignite.IgniteCheckedException: Failed to read WAL record at 
> position: 28310114 size: -1
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.handleRecordException(AbstractWalRecordsIterator.java:292)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:258)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advance(AbstractWalRecordsIterator.java:154)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.SingleSegmentLogicalRecordsIterator.advance(SingleSegmentLogicalRecordsIterator.java:119)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:123)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:52)
> at 
> org.apache.ignite.internal.util.GridCloseableIteratorAdapter.nextX(GridCloseableIteratorAdapter.java:41)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.compressSegmentToFile(FileWriteAheadLogManager.java:2039)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body0(FileWriteAheadLogManager.java:1974)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body(FileWriteAheadLogManager.java:1950)
> at 
> org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: class org.apache.ignite.IgniteCheckedException: Failed to read WAL 
> record at position: 28310114 size: -1
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:394)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer.readRecord(RecordV2Serializer.java:235)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:243)
> ... 10 more
> Caused by: java.nio.channels.ClosedByInterruptException
> at 
> java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
> at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:164)
> at 
> org.apache.ignite.internal.processors.cache.persistence.file.RandomAccessFileIO.read(RandomAccessFileIO.java:58)
> at 
> org.apache.ignite.internal.processors.cache.persistence.file.FileIODecorator.read(FileIODecorator.java:51)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.SimpleFileInput.ensure(SimpleFileInput.java:119)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.ensure(FileInput.java:89)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.readFully(FileInput.java:152)
> at 
> 

[jira] [Created] (IGNITE-10928) After huge load on cluster and restart with walCompactionEnabled=True errors on log

2019-01-14 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10928:
---

 Summary: After huge load on cluster and restart with 
walCompactionEnabled=True errors on log
 Key: IGNITE-10928
 URL: https://issues.apache.org/jira/browse/IGNITE-10928
 Project: Ignite
  Issue Type: Bug
Reporter: ARomantsov



{code:java}







{code}



{code:java}
[15:30:56,809][INFO][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
 Stopping WAL iteration due to an exception: Failed to read WAL record at 
position: 28310114 size: -1, ptr=FileWALPointer [idx=35, fileOff=28310114, 
len=0]
[15:30:56,811][INFO][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
 Stopping WAL iteration due to an exception: Failed to read WAL record at 
position: 28303753 size: -1, ptr=FileWALPointer [idx=36, fileOff=28303753, 
len=0]
[15:30:56,811][SEVERE][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
 Compression of WAL segment [idx=35] was skipped due to unexpected error
class org.apache.ignite.IgniteCheckedException: Failed to read WAL record at 
position: 28310114 size: -1
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.handleRecordException(AbstractWalRecordsIterator.java:292)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:258)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advance(AbstractWalRecordsIterator.java:154)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.SingleSegmentLogicalRecordsIterator.advance(SingleSegmentLogicalRecordsIterator.java:119)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:123)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:52)
at 
org.apache.ignite.internal.util.GridCloseableIteratorAdapter.nextX(GridCloseableIteratorAdapter.java:41)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.compressSegmentToFile(FileWriteAheadLogManager.java:2039)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body0(FileWriteAheadLogManager.java:1974)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body(FileWriteAheadLogManager.java:1950)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
at java.lang.Thread.run(Thread.java:748)
Caused by: class org.apache.ignite.IgniteCheckedException: Failed to read WAL 
record at position: 28310114 size: -1
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:394)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer.readRecord(RecordV2Serializer.java:235)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:243)
... 10 more
Caused by: java.nio.channels.ClosedByInterruptException
at 
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:164)
at 
org.apache.ignite.internal.processors.cache.persistence.file.RandomAccessFileIO.read(RandomAccessFileIO.java:58)
at 
org.apache.ignite.internal.processors.cache.persistence.file.FileIODecorator.read(FileIODecorator.java:51)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.io.SimpleFileInput.ensure(SimpleFileInput.java:119)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.ensure(FileInput.java:89)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.readFully(FileInput.java:152)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer$2.readWithHeaders(RecordV2Serializer.java:149)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:373)
... 12 more
Suppressed: class 
org.apache.ignite.internal.processors.cache.persistence.wal.crc.IgniteDataIntegrityViolationException:
 val: 1241861030 writtenCrc: 20736
at 
org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.close(FileInput.java:106)
 

[jira] [Updated] (IGNITE-10928) After huge load on cluster and restart with walCompactionEnabled=True errors on log

2019-01-14 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10928:

Priority: Critical  (was: Major)

> After huge load on cluster and restart with walCompactionEnabled=True errors 
> on log
> ---
>
> Key: IGNITE-10928
> URL: https://issues.apache.org/jira/browse/IGNITE-10928
> Project: Ignite
>  Issue Type: Bug
>Reporter: ARomantsov
>Priority: Critical
>
> {code:java}
> 
>  class="org.apache.ignite.configuration.DataRegionConfiguration">
> 
> 
> 
> 
> 
> {code}
> {code:java}
> [15:30:56,809][INFO][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
>  Stopping WAL iteration due to an exception: Failed to read WAL record at 
> position: 28310114 size: -1, ptr=FileWALPointer [idx=35, fileOff=28310114, 
> len=0]
> [15:30:56,811][INFO][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
>  Stopping WAL iteration due to an exception: Failed to read WAL record at 
> position: 28303753 size: -1, ptr=FileWALPointer [idx=36, fileOff=28303753, 
> len=0]
> [15:30:56,811][SEVERE][wal-file-compressor-%null%-1-#68][FileWriteAheadLogManager]
>  Compression of WAL segment [idx=35] was skipped due to unexpected error
> class org.apache.ignite.IgniteCheckedException: Failed to read WAL record at 
> position: 28310114 size: -1
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.handleRecordException(AbstractWalRecordsIterator.java:292)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:258)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advance(AbstractWalRecordsIterator.java:154)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.SingleSegmentLogicalRecordsIterator.advance(SingleSegmentLogicalRecordsIterator.java:119)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:123)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:52)
> at 
> org.apache.ignite.internal.util.GridCloseableIteratorAdapter.nextX(GridCloseableIteratorAdapter.java:41)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.compressSegmentToFile(FileWriteAheadLogManager.java:2039)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body0(FileWriteAheadLogManager.java:1974)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body(FileWriteAheadLogManager.java:1950)
> at 
> org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: class org.apache.ignite.IgniteCheckedException: Failed to read WAL 
> record at position: 28310114 size: -1
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:394)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer.readRecord(RecordV2Serializer.java:235)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:243)
> ... 10 more
> Caused by: java.nio.channels.ClosedByInterruptException
> at 
> java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
> at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:164)
> at 
> org.apache.ignite.internal.processors.cache.persistence.file.RandomAccessFileIO.read(RandomAccessFileIO.java:58)
> at 
> org.apache.ignite.internal.processors.cache.persistence.file.FileIODecorator.read(FileIODecorator.java:51)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.SimpleFileInput.ensure(SimpleFileInput.java:119)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.ensure(FileInput.java:89)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.readFully(FileInput.java:152)
> at 
> org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer$2.readWithHeaders(RecordV2Serializer.java:149)
> 

[jira] [Created] (IGNITE-10911) Need to enforce truststore key when control utility connect to cluster with SSL enabled

2019-01-13 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10911:
---

 Summary: Need to enforce truststore key when control utility 
connect to cluster with SSL enabled
 Key: IGNITE-10911
 URL: https://issues.apache.org/jira/browse/IGNITE-10911
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.5
Reporter: ARomantsov
 Fix For: 2.8


I use the next setting in my app
{code:java}
































{code}

And can get the result of command - control.bat --baseline 
--keystore..\server.jks --keystore-password * without set up truststore




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10896) Add ability to use more than one key with control.sh --cache idle_verify

2019-01-11 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10896:

Description: 
Now I can use only one of next options
1) --exclude-caches cache1,...,cacheN
2) --cache-filter ALL|SYSTEM|PERSISTENT|NOT_PERSISTENT
3) cache1,...,cacheN

I suppose that using 1 and 2 or 2 and 3 make this command more flexiable

  was:
Now I can use only one of next options
1) --exclude-caches cache1,...,cacheN or 
2) --cache-filter ALL|SYSTEM|PERSISTENT|NOT_PERSISTENT
3) cache1,...,cacheN

I suppose that using 1 and 2 or 2 and 3 make this command more flexiable


> Add ability to use more than one key with control.sh --cache idle_verify
> 
>
> Key: IGNITE-10896
> URL: https://issues.apache.org/jira/browse/IGNITE-10896
> Project: Ignite
>  Issue Type: Improvement
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> Now I can use only one of next options
> 1) --exclude-caches cache1,...,cacheN
> 2) --cache-filter ALL|SYSTEM|PERSISTENT|NOT_PERSISTENT
> 3) cache1,...,cacheN
> I suppose that using 1 and 2 or 2 and 3 make this command more flexiable



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10896) Add ability to use more than one key with control.sh --cache idle_verify

2019-01-11 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10896:

Fix Version/s: 2.8

> Add ability to use more than one key with control.sh --cache idle_verify
> 
>
> Key: IGNITE-10896
> URL: https://issues.apache.org/jira/browse/IGNITE-10896
> Project: Ignite
>  Issue Type: Improvement
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> Now I can use only one of next options
> 1) --exclude-caches cache1,...,cacheN or 
> 2) --cache-filter ALL|SYSTEM|PERSISTENT|NOT_PERSISTENT
> 3) cache1,...,cacheN
> I suppose that using 1 and 2 or 2 and 3 make this command more flexiable



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10896) Add ability to use more than one key with control.sh --cache idle_verify

2019-01-11 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10896:
---

 Summary: Add ability to use more than one key with control.sh 
--cache idle_verify
 Key: IGNITE-10896
 URL: https://issues.apache.org/jira/browse/IGNITE-10896
 Project: Ignite
  Issue Type: Improvement
Reporter: ARomantsov


Now I can use only one of next options
1) --exclude-caches cache1,...,cacheN or 
2) --cache-filter ALL|SYSTEM|PERSISTENT|NOT_PERSISTENT
3) cache1,...,cacheN

I suppose that using 1 and 2 or 2 and 3 make this command more flexiable



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10848) Add abbitlity to set default log directory by JVM key and set it by default to /var/log in *nix system

2018-12-29 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10848:
---

 Summary: Add abbitlity to set default log directory by JVM key and 
set it by default to /var/log in *nix system
 Key: IGNITE-10848
 URL: https://issues.apache.org/jira/browse/IGNITE-10848
 Project: Ignite
  Issue Type: Improvement
Reporter: ARomantsov






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10825) After node restart and and new node to BLT due load - some partition inconsistent

2018-12-26 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10825:

Description: 

{code:java}
14:12:20 [14:12:20][:573 :252] idle_verify check has finished, found 2 conflict 
partitions: [counterConflicts=1, hashConflicts=1]
14:12:20 [14:12:20][:573 :252] Update counter conflicts:
14:12:20 [14:12:20][:573 :252] Conflict partition: PartitionKeyV2 
[grpId=374280887, grpName=cache_group_4, partId=115]
14:12:20 [14:12:20][:573 :252] Partition instances: 

  [PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_5, 
updateCntr=10, size=2, partHash=-979021948], 

   PartitionHashRecordV2 [isPrimary=true, consistentId=node_1_2, 
updateCntr=11, size=2, partHash=-731597536], 

   PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_1, 
updateCntr=11, size=2, partHash=-731597536], 

   PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_3, 
updateCntr=11, size=2, partHash=-731597536], 

   PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_6, 
updateCntr=11, size=2, partHash=-731597536], 

   PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_4, 
updateCntr=11, size=2, partHash=-731597536], 

   PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_10001, 
updateCntr=11, size=2, partHash=-731597536]]
14:12:20 [14:12:20][:573 :252] Hash conflicts:
14:12:20 [14:12:20][:573 :252] Conflict partition: PartitionKeyV2 
[grpId=374280887, grpName=cache_group_4, partId=115]
14:12:20 [14:12:20][:573 :252] Partition instances: 

[PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_5, 
updateCntr=10, size=2, partHash=-979021948], 

PartitionHashRecordV2 [isPrimary=true, consistentId=node_1_2, 
updateCntr=11, size=2, partHash=-731597536], 

PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_1, 
updateCntr=11, size=2, partHash=-731597536], 

PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_3, 
updateCntr=11, size=2, partHash=-731597536],

PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_6, 
updateCntr=11, size=2, partHash=-731597536], 

PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_4, 
updateCntr=11, size=2, partHash=-731597536], 

PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_10001, 
updateCntr=11, size=2, partHash=-731597536]]
{code}


> After node restart and and new node to BLT due load - some partition 
> inconsistent
> -
>
> Key: IGNITE-10825
> URL: https://issues.apache.org/jira/browse/IGNITE-10825
> Project: Ignite
>  Issue Type: Bug
>  Components: data structures
>Affects Versions: 2.8
>Reporter: ARomantsov
>Priority: Critical
>
> {code:java}
> 14:12:20 [14:12:20][:573 :252] idle_verify check has finished, found 2 
> conflict partitions: [counterConflicts=1, hashConflicts=1]
> 14:12:20 [14:12:20][:573 :252] Update counter conflicts:
> 14:12:20 [14:12:20][:573 :252] Conflict partition: PartitionKeyV2 
> [grpId=374280887, grpName=cache_group_4, partId=115]
> 14:12:20 [14:12:20][:573 :252] Partition instances: 
>   
>   [PartitionHashRecordV2 [isPrimary=false, consistentId=node_1_5, 
> updateCntr=10, size=2, partHash=-979021948], 
>   
>PartitionHashRecordV2 [isPrimary=true, consistentId=node_1_2, 
> updateCntr=11, size=2, partHash=-731597536], 
>   
>

[jira] [Created] (IGNITE-10825) After node restart and and new node to BLT due load - some partition inconsistent

2018-12-26 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10825:
---

 Summary: After node restart and and new node to BLT due load - 
some partition inconsistent
 Key: IGNITE-10825
 URL: https://issues.apache.org/jira/browse/IGNITE-10825
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.8
Reporter: ARomantsov






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10772) If version look like X.X.X.X rest version return X.X.X-X

2018-12-20 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10772:
---

 Summary: If version look like X.X.X.X rest version return X.X.X-X
 Key: IGNITE-10772
 URL: https://issues.apache.org/jira/browse/IGNITE-10772
 Project: Ignite
  Issue Type: Bug
  Components: rest
Affects Versions: 2.8
Reporter: ARomantsov
 Fix For: 2.8


Test URL - http://localhost:8080/ignite?cmd=version
{
  "result": {
"error": null,
"response": "X.X.X-X",
"sessionToken": "D372FC2DD4A24603AC39CC92C6B132EC",
"successStatus": 0
  },
  "status": "OK"
}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10744) class org.apache.ignite.IgniteCheckedException: Failed to read WAL record at position: 5237375 size: -1

2018-12-19 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10744:
---

 Summary: class org.apache.ignite.IgniteCheckedException: Failed to 
read WAL record at position: 5237375 size: -1
 Key: IGNITE-10744
 URL: https://issues.apache.org/jira/browse/IGNITE-10744
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.8
Reporter: ARomantsov
 Fix For: 2.8


Scenario:

-Start 4 nodes with disc storage
-Little load
-Deactivate
-Activate

Got next message in log


{code:java}
[20:54:42,746][INFO][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
 Stopping WAL iteration due to an exception: Failed to read WAL record at 
position: 5237375 size: -1, ptr=FileWALPointer [idx=4, fileOff=5237375, len=0]
[20:54:42,747][SEVERE][wal-file-compressor-%null%-3-#70][FileWriteAheadLogManager]
 Compression of WAL segment [idx=4] was skipped due to unexpected error
class org.apache.ignite.IgniteCheckedException: Failed to read WAL record at 
position: 5237375 size: -1
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.handleRecordException(AbstractWalRecordsIterator.java:294)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:258)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advance(AbstractWalRecordsIterator.java:154)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.SingleSegmentLogicalRecordsIterator.advance(SingleSegmentLogicalRecordsIterator.java:119)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:123)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.onNext(AbstractWalRecordsIterator.java:52)
at 
org.apache.ignite.internal.util.GridCloseableIteratorAdapter.nextX(GridCloseableIteratorAdapter.java:41)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.compressSegmentToFile(FileWriteAheadLogManager.java:2186)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body0(FileWriteAheadLogManager.java:2111)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileCompressorWorker.body(FileWriteAheadLogManager.java:2081)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:120)
at java.lang.Thread.run(Thread.java:748)
Caused by: class org.apache.ignite.IgniteCheckedException: Failed to read WAL 
record at position: 5237375 size: -1
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:394)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer.readRecord(RecordV2Serializer.java:235)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.AbstractWalRecordsIterator.advanceRecord(AbstractWalRecordsIterator.java:243)
... 10 more
Caused by: java.nio.channels.ClosedByInterruptException
at 
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:164)
at 
org.apache.ignite.internal.processors.cache.persistence.file.RandomAccessFileIO.read(RandomAccessFileIO.java:58)
at 
org.apache.ignite.internal.processors.cache.persistence.file.FileIODecorator.read(FileIODecorator.java:51)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.io.SimpleFileInput.ensure(SimpleFileInput.java:119)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.ensure(FileInput.java:89)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.skipBytes(FileInput.java:130)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV2Serializer$2.readWithHeaders(RecordV2Serializer.java:130)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:373)
... 12 more
Suppressed: class 
org.apache.ignite.internal.processors.cache.persistence.wal.crc.IgniteDataIntegrityViolationException:
 val: 505117685 writtenCrc: 0
at 
org.apache.ignite.internal.processors.cache.persistence.wal.io.FileInput$Crc32CheckingFileInput.close(FileInput.java:106)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.serializer.RecordV1Serializer.readWithCrc(RecordV1Serializer.java:380)
... 12 more
{code}




--
This 

[jira] [Commented] (IGNITE-8225) Add a command to control script to print current topology version

2018-12-19 Thread ARomantsov (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8225?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16724986#comment-16724986
 ] 

ARomantsov commented on IGNITE-8225:


[~ruchirc], hello
Can you have any objection if i take this ticker.
I'm going to do it on Friday 12/21/2018

> Add a command to control script to print current topology version
> -
>
> Key: IGNITE-8225
> URL: https://issues.apache.org/jira/browse/IGNITE-8225
> Project: Ignite
>  Issue Type: Improvement
>Reporter: Alexey Goncharuk
>Assignee: ruchir choudhry
>Priority: Critical
> Fix For: 2.8
>
>
> The command should be {{./control.sh --topology}} and should print a short 
> summary about the current topology (topology version, number of client nodes, 
> number of server nodes, baseline topology information)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (IGNITE-10279) Control.sh utility unify options naming format

2018-12-14 Thread ARomantsov (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-10279?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16721367#comment-16721367
 ] 

ARomantsov commented on IGNITE-10279:
-

Can you please also fix help of control.sh

I suggest something like

{code:java}
Contol.sh is used to execute admin commands on cluster or get common cluster 
info. The command has the following syntax:
  control.sh [--host HOST_OR_IP] [--port PORT] [--user USER] [--password 
PASSWORD] [--ping-interval PING_INTERVAL] [--ping-timeout PING_TIMEOUT] 
[--ssl_enabled] [--ssl_protocol SSL_PROTOCOL[, SSL_PROTOCOL_2, ...]] 
[--ssl_algorithm SSL_ALGORITHM] [--ssl_cipher_suites SSL_CIPHER_1[, 
SSL_CIPHER_2, ...]] [--ssl_key_store_type SSL_KEY_STORE_TYPE] 
[--ssl_key_store_path PATH] [--ssl_key_store_password PASSWORD] 
[--ssl_truststore_type SSL_TRUST_STORE_TYPE] [--ssl_truststore_path PATH] 
[--ssl_truststore_password PASSWORD] [command] 
This utility can do the following commands:
  Activate cluster:
control.sh  --activate
  Deactivate cluster:
control.sh  --deactivate [--yes]
  Print current cluster state:
control.sh  --state
  Print cluster baseline topology:
control.sh  --baseline
  Add nodes into baseline topology:
control.sh  --baseline add consistentId1[,consistentId2,,consistentIdN] 
[--yes]
  Remove nodes from baseline topology:
control.sh  --baseline remove 
consistentId1[,consistentId2,,consistentIdN] [--yes]
  Set baseline topology:
control.sh  --baseline set consistentId1[,consistentId2,,consistentIdN] 
[--yes]
  Set baseline topology based on version:
control.sh  --baseline version topologyVersion [--yes]
  List or kill transactions:
control.sh  --tx [xid XID] [minDuration SECONDS] [minSize SIZE] [label 
PATTERN_REGEX] [servers|clients] [nodes 
consistentId1[,consistentId2,,consistentIdN]] [limit NUMBER] [order 
DURATION|SIZE|START_TIME] [kill] [--yes]
  View caches information in a cluster. For more details type:
control.sh --cache help
By default commands affecting the cluster require interactive confirmation.
Use --yes option to disable it.
Default values:
HOST_OR_IP=127.0.0.1
PORT=11211
PING_INTERVAL=5000
PING_TIMEOUT=3
SSL_PROTOCOL=TLS
SSL_ALGORITHM=SunX509
SSL_KEY_STORE_TYPE=JKS
SSL_TRUST_STORE_TYPE=JKS
Exit codes:
0 - successful execution.
1 - invalid arguments.
2 - connection failed.
3 - authentication failed.
4 - unexpected error.
{code}


> Control.sh utility unify options naming format
> --
>
> Key: IGNITE-10279
> URL: https://issues.apache.org/jira/browse/IGNITE-10279
> Project: Ignite
>  Issue Type: Improvement
>Affects Versions: 2.6
>Reporter: Sergey Antonov
>Assignee: Sergey Antonov
>Priority: Major
> Fix For: 2.8
>
>
> Now we have options in several styles:
> {noformat}
> --ping-interval 
> {noformat}
> {noformat}
> --skipZeros
> {noformat}
> I think, we must unify options naming format and we should use linux like 
> format, i.e. {{--word1-word2}}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10669) NPE in freelist.PagesList.findTailIndex

2018-12-13 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10669:

Description: 
Run node with 1 cache and put to it.
Kill node and try run back - it broken on start

{code:java}
[22:40:10,916][INFO][main][GridCacheDatabaseSharedManager] Applying lost cache 
updates since last checkpoint record [lastMarked=FileWALPointer [idx=2, 
fileOff=14706, len=21409], 
lastCheckpointId=2f9202e9-c9d7-47ca-9dcc-299a959bb2e0]
[22:40:10,922][SEVERE][main][IgniteKernal] Exception during start processors, 
node will be stopped and close connections
java.lang.NullPointerException
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.findTailIndex(PagesList.java:502)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.updateTail(PagesList.java:458)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.mergeNoNext(PagesList.java:1330)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.removeDataPage(PagesList.java:1281)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList$RemoveRowHandler.run(AbstractFreeList.java:305)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList$RemoveRowHandler.run(AbstractFreeList.java:261)
at 
org.apache.ignite.internal.processors.cache.persistence.tree.util.PageHandler.writePage(PageHandler.java:279)
at 
org.apache.ignite.internal.processors.cache.persistence.DataStructure.write(DataStructure.java:256)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList.removeDataRowByLink(AbstractFreeList.java:571)
at 
org.apache.ignite.internal.processors.cache.persistence.metastorage.MetastorageRowStore.removeRow(MetastorageRowStore.java:57)
at 
org.apache.ignite.internal.processors.cache.persistence.metastorage.MetaStorage.putData(MetaStorage.java:253)
at 
org.apache.ignite.internal.processors.cache.persistence.metastorage.MetaStorage.applyUpdate(MetaStorage.java:492)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager.applyLogicalUpdates(GridCacheDatabaseSharedManager.java:2420)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager.startMemoryRestore(GridCacheDatabaseSharedManager.java:1909)
at org.apache.ignite.internal.IgniteKernal.start(IgniteKernal.java:1056)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance.start0(IgnitionEx.java:2040)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance.start(IgnitionEx.java:1732)
at org.apache.ignite.internal.IgnitionEx.start0(IgnitionEx.java:1158)
at 
org.apache.ignite.internal.IgnitionEx.startConfigurations(IgnitionEx.java:1076)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:962)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:861)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:731)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:700)
at org.apache.ignite.Ignition.start(Ignition.java:348)
at 
org.apache.ignite.startup.cmdline.CommandLineStartup.main(CommandLineStartup.java:301)
[22:40:10,922][SEVERE][main][IgniteKernal] Got exception while starting (will 
rollback startup routine).
java.lang.NullPointerException
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.findTailIndex(PagesList.java:502)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.updateTail(PagesList.java:458)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.mergeNoNext(PagesList.java:1330)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.removeDataPage(PagesList.java:1281)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList$RemoveRowHandler.run(AbstractFreeList.java:305)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList$RemoveRowHandler.run(AbstractFreeList.java:261)
at 
org.apache.ignite.internal.processors.cache.persistence.tree.util.PageHandler.writePage(PageHandler.java:279)
at 
org.apache.ignite.internal.processors.cache.persistence.DataStructure.write(DataStructure.java:256)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList.removeDataRowByLink(AbstractFreeList.java:571)
at 
org.apache.ignite.internal.processors.cache.persistence.metastorage.MetastorageRowStore.removeRow(MetastorageRowStore.java:57)
at 

[jira] [Created] (IGNITE-10669) NPE in freelist.PagesList.findTailIndex

2018-12-13 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10669:
---

 Summary: NPE in freelist.PagesList.findTailIndex
 Key: IGNITE-10669
 URL: https://issues.apache.org/jira/browse/IGNITE-10669
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.7
 Environment: Windows
Reporter: ARomantsov
 Fix For: 2.8



{code:java}
[22:40:10,916][INFO][main][GridCacheDatabaseSharedManager] Applying lost cache 
updates since last checkpoint record [lastMarked=FileWALPointer [idx=2, 
fileOff=14706, len=21409], 
lastCheckpointId=2f9202e9-c9d7-47ca-9dcc-299a959bb2e0]
[22:40:10,922][SEVERE][main][IgniteKernal] Exception during start processors, 
node will be stopped and close connections
java.lang.NullPointerException
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.findTailIndex(PagesList.java:502)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.updateTail(PagesList.java:458)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.mergeNoNext(PagesList.java:1330)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.removeDataPage(PagesList.java:1281)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList$RemoveRowHandler.run(AbstractFreeList.java:305)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList$RemoveRowHandler.run(AbstractFreeList.java:261)
at 
org.apache.ignite.internal.processors.cache.persistence.tree.util.PageHandler.writePage(PageHandler.java:279)
at 
org.apache.ignite.internal.processors.cache.persistence.DataStructure.write(DataStructure.java:256)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList.removeDataRowByLink(AbstractFreeList.java:571)
at 
org.apache.ignite.internal.processors.cache.persistence.metastorage.MetastorageRowStore.removeRow(MetastorageRowStore.java:57)
at 
org.apache.ignite.internal.processors.cache.persistence.metastorage.MetaStorage.putData(MetaStorage.java:253)
at 
org.apache.ignite.internal.processors.cache.persistence.metastorage.MetaStorage.applyUpdate(MetaStorage.java:492)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager.applyLogicalUpdates(GridCacheDatabaseSharedManager.java:2420)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager.startMemoryRestore(GridCacheDatabaseSharedManager.java:1909)
at org.apache.ignite.internal.IgniteKernal.start(IgniteKernal.java:1056)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance.start0(IgnitionEx.java:2040)
at 
org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance.start(IgnitionEx.java:1732)
at org.apache.ignite.internal.IgnitionEx.start0(IgnitionEx.java:1158)
at 
org.apache.ignite.internal.IgnitionEx.startConfigurations(IgnitionEx.java:1076)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:962)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:861)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:731)
at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:700)
at org.apache.ignite.Ignition.start(Ignition.java:348)
at 
org.apache.ignite.startup.cmdline.CommandLineStartup.main(CommandLineStartup.java:301)
[22:40:10,922][SEVERE][main][IgniteKernal] Got exception while starting (will 
rollback startup routine).
java.lang.NullPointerException
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.findTailIndex(PagesList.java:502)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.updateTail(PagesList.java:458)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.mergeNoNext(PagesList.java:1330)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.PagesList.removeDataPage(PagesList.java:1281)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList$RemoveRowHandler.run(AbstractFreeList.java:305)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList$RemoveRowHandler.run(AbstractFreeList.java:261)
at 
org.apache.ignite.internal.processors.cache.persistence.tree.util.PageHandler.writePage(PageHandler.java:279)
at 
org.apache.ignite.internal.processors.cache.persistence.DataStructure.write(DataStructure.java:256)
at 
org.apache.ignite.internal.processors.cache.persistence.freelist.AbstractFreeList.removeDataRowByLink(AbstractFreeList.java:571)
at 

[jira] [Updated] (IGNITE-10664) need to make identical the behavior of control.sh and control.bat

2018-12-12 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10664:

Description: 
Now control.bat contain next line: 
if not "%NO_PAUSE%" == "1" pause 
and after execution it wait on - Press any key to continue...

on the other hand - control.sh not contain any pause

to use automation with ignite the second option is better

  was:
Now control.bat contain next line: 
if not "%NO_PAUSE%" == "1" pause 
and after execution it wait on - Press any key to continue...

on the other hand - control.sh not contain any pause


>  need to make identical the behavior of control.sh and control.bat
> --
>
> Key: IGNITE-10664
> URL: https://issues.apache.org/jira/browse/IGNITE-10664
> Project: Ignite
>  Issue Type: Improvement
>  Components: clients
>Affects Versions: 2.7
> Environment: windows, nix
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> Now control.bat contain next line: 
> if not "%NO_PAUSE%" == "1" pause 
> and after execution it wait on - Press any key to continue...
> on the other hand - control.sh not contain any pause
> to use automation with ignite the second option is better



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10664) need to make identical the behavior of control.sh and control.bat

2018-12-12 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10664:

Priority: Major  (was: Minor)

>  need to make identical the behavior of control.sh and control.bat
> --
>
> Key: IGNITE-10664
> URL: https://issues.apache.org/jira/browse/IGNITE-10664
> Project: Ignite
>  Issue Type: Improvement
>  Components: clients
>Affects Versions: 2.7
> Environment: windows, nix
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> Now control.bat contain next line: 
> if not "%NO_PAUSE%" == "1" pause 
> and after execution it wait on - Press any key to continue...
> on the other hand - control.sh not contain any pause



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10664) need to make identical the behavior of control.sh and control.bat

2018-12-12 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10664:

Description: 
Now control.bat contain next line: 
if not "%NO_PAUSE%" == "1" pause 
and after execution it wait on - Press any key to continue...

on the other hand - control.sh not contain any pause

  was:
Now control.bat contain next line: 
if not "%NO_PAUSE%" == "1" pause 
and after execution it wait  press any button.

on the other hand - control.sh not contain any pause


>  need to make identical the behavior of control.sh and control.bat
> --
>
> Key: IGNITE-10664
> URL: https://issues.apache.org/jira/browse/IGNITE-10664
> Project: Ignite
>  Issue Type: Bug
>  Components: clients
>Affects Versions: 2.7
> Environment: windows, nix
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> Now control.bat contain next line: 
> if not "%NO_PAUSE%" == "1" pause 
> and after execution it wait on - Press any key to continue...
> on the other hand - control.sh not contain any pause



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10664) need to make identical the behavior of control.sh and control.bat

2018-12-12 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10664:

Priority: Minor  (was: Major)

>  need to make identical the behavior of control.sh and control.bat
> --
>
> Key: IGNITE-10664
> URL: https://issues.apache.org/jira/browse/IGNITE-10664
> Project: Ignite
>  Issue Type: Bug
>  Components: clients
>Affects Versions: 2.7
> Environment: windows, nix
>Reporter: ARomantsov
>Priority: Minor
> Fix For: 2.8
>
>
> Now control.bat contain next line: 
> if not "%NO_PAUSE%" == "1" pause 
> and after execution it wait on - Press any key to continue...
> on the other hand - control.sh not contain any pause



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10664) need to make identical the behavior of control.sh and control.bat

2018-12-12 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10664:

Issue Type: Improvement  (was: Bug)

>  need to make identical the behavior of control.sh and control.bat
> --
>
> Key: IGNITE-10664
> URL: https://issues.apache.org/jira/browse/IGNITE-10664
> Project: Ignite
>  Issue Type: Improvement
>  Components: clients
>Affects Versions: 2.7
> Environment: windows, nix
>Reporter: ARomantsov
>Priority: Minor
> Fix For: 2.8
>
>
> Now control.bat contain next line: 
> if not "%NO_PAUSE%" == "1" pause 
> and after execution it wait on - Press any key to continue...
> on the other hand - control.sh not contain any pause



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10664) need to make identical the behavior of control.sh and control.bat

2018-12-12 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10664:

Description: 
Now control.bat contain next line: 
if not "%NO_PAUSE%" == "1" pause 
and after execution it wait  press any button.

on the other hand - control.sh not contain any pause

  was:
Now control.bat contain next line: 
if not "%NO_PAUSE%" == "1" pause 
and after execution it ready to press button.

on the other hand - control.sh not contain any pause


>  need to make identical the behavior of control.sh and control.bat
> --
>
> Key: IGNITE-10664
> URL: https://issues.apache.org/jira/browse/IGNITE-10664
> Project: Ignite
>  Issue Type: Bug
>  Components: clients
>Affects Versions: 2.7
> Environment: windows, nix
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> Now control.bat contain next line: 
> if not "%NO_PAUSE%" == "1" pause 
> and after execution it wait  press any button.
> on the other hand - control.sh not contain any pause



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10664) need to make identical the behavior of control.sh and control.bat

2018-12-12 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10664:
---

 Summary:  need to make identical the behavior of control.sh and 
control.bat
 Key: IGNITE-10664
 URL: https://issues.apache.org/jira/browse/IGNITE-10664
 Project: Ignite
  Issue Type: Bug
  Components: clients
Affects Versions: 2.7
 Environment: windows, nix
Reporter: ARomantsov
 Fix For: 2.8


Now control.bat contain next line: 
if not "%NO_PAUSE%" == "1" pause 
and after execution it ready to press button.

on the other hand - control.sh not contain any pause



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (IGNITE-9356) Ignite rest command http://localhost:8080/ignite?cmd=log=n=m return more line in linux than windows

2018-12-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9356?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov closed IGNITE-9356.
--

> Ignite rest command http://localhost:8080/ignite?cmd=log=n=m return 
> more line in linux than windows  
> -
>
> Key: IGNITE-9356
> URL: https://issues.apache.org/jira/browse/IGNITE-9356
> Project: Ignite
>  Issue Type: Improvement
>  Components: rest
>Affects Versions: 2.5
> Environment: Centos/ Windows10
>Reporter: ARomantsov
>Priority: Minor
> Fix For: 2.8
>
>
> I run cluster in diffrent configuration (centos and windows 10) and notice 
> that log command return diffrent count of rows in same from and to
> Windows rest return 1 less rows



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9356) Ignite rest command http://localhost:8080/ignite?cmd=log=n=m return more line in linux than windows

2018-12-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9356?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9356:
---
Ignite Flags:   (was: Docs Required)

> Ignite rest command http://localhost:8080/ignite?cmd=log=n=m return 
> more line in linux than windows  
> -
>
> Key: IGNITE-9356
> URL: https://issues.apache.org/jira/browse/IGNITE-9356
> Project: Ignite
>  Issue Type: Improvement
>  Components: rest
>Affects Versions: 2.5
> Environment: Centos/ Windows10
>Reporter: ARomantsov
>Priority: Minor
> Fix For: 2.8
>
>
> I run cluster in diffrent configuration (centos and windows 10) and notice 
> that log command return diffrent count of rows in same from and to
> Windows rest return 1 less rows



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (IGNITE-9356) Ignite rest command http://localhost:8080/ignite?cmd=log=n=m return more line in linux than windows

2018-12-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9356?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov resolved IGNITE-9356.

Resolution: Cannot Reproduce

> Ignite rest command http://localhost:8080/ignite?cmd=log=n=m return 
> more line in linux than windows  
> -
>
> Key: IGNITE-9356
> URL: https://issues.apache.org/jira/browse/IGNITE-9356
> Project: Ignite
>  Issue Type: Improvement
>  Components: rest
>Affects Versions: 2.5
> Environment: Centos/ Windows10
>Reporter: ARomantsov
>Priority: Minor
> Fix For: 2.8
>
>
> I run cluster in diffrent configuration (centos and windows 10) and notice 
> that log command return diffrent count of rows in same from and to
> Windows rest return 1 less rows



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10504) If client have cache resource with not configurate data region it stop by handler

2018-12-03 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10504:

Description: 

{code:java}
[16:02:08,932][SEVERE][exchange-worker-#58][] Critical system error detected. 
Will be handled accordingly to configured handler 
[hnd=StopNodeOrHaltFailureHandler [tryStop=false, timeout=0, 
super=AbstractFailureHandler [ignoredFailureTyp
es=SingletonSet [SYSTEM_WORKER_BLOCKED]]], failureCtx=FailureContext 
[type=SYSTEM_WORKER_TERMINATION, err=class o.a.i.IgniteCheckedException: 
Requested DataRegion is not configured: region-name2]]
class org.apache.ignite.IgniteCheckedException: Requested DataRegion is not 
configured: region-name2
at 
org.apache.ignite.internal.processors.cache.persistence.IgniteCacheDatabaseSharedManager.dataRegion(IgniteCacheDatabaseSharedManager.java:729)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.startCacheGroup(GridCacheProcessor.java:2641)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.lambda$getOrCreateCacheGroupContext$8(GridCacheProcessor.java:2415)
at 
org.apache.ignite.internal.util.InitializationProtector.protect(InitializationProtector.java:60)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.getOrCreateCacheGroupContext(GridCacheProcessor.java:2412)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.prepareCacheContext(GridCacheProcessor.java:2263)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.lambda$null$fd62dedb$1(GridCacheProcessor.java:2110)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.lambda$prepareStartCaches$4(GridCacheProcessor.java:2033)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.lambda$prepareStartCaches$937cbe24$1(GridCacheProcessor.java:2107)
at 
org.apache.ignite.internal.util.IgniteUtils.doInParallel(IgniteUtils.java:10891)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.prepareStartCaches(GridCacheProcessor.java:2102)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.prepareStartCaches(GridCacheProcessor.java:2032)
at 
org.apache.ignite.internal.processors.cache.GridCacheProcessor.startCachesOnLocalJoin(GridCacheProcessor.java:1978)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.initCachesOnLocalJoin(GridDhtPartitionsExchangeFuture.java:934)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.init(GridDhtPartitionsExchangeFuture.java:796)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$ExchangeWorker.body0(GridCachePartitionExchangeManager.java:2904)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$ExchangeWorker.body(GridCachePartitionExchangeManager.java:2761)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:120)
at java.lang.Thread.run(Thread.java:748)
{code}


> If client have cache resource with not configurate data region it stop by 
> handler
> -
>
> Key: IGNITE-10504
> URL: https://issues.apache.org/jira/browse/IGNITE-10504
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Minor
> Fix For: 2.8
>
>
> {code:java}
> [16:02:08,932][SEVERE][exchange-worker-#58][] Critical system error detected. 
> Will be handled accordingly to configured handler 
> [hnd=StopNodeOrHaltFailureHandler [tryStop=false, timeout=0, 
> super=AbstractFailureHandler [ignoredFailureTyp
> es=SingletonSet [SYSTEM_WORKER_BLOCKED]]], failureCtx=FailureContext 
> [type=SYSTEM_WORKER_TERMINATION, err=class o.a.i.IgniteCheckedException: 
> Requested DataRegion is not configured: region-name2]]
> class org.apache.ignite.IgniteCheckedException: Requested DataRegion is not 
> configured: region-name2
> at 
> org.apache.ignite.internal.processors.cache.persistence.IgniteCacheDatabaseSharedManager.dataRegion(IgniteCacheDatabaseSharedManager.java:729)
> at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.startCacheGroup(GridCacheProcessor.java:2641)
> at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.lambda$getOrCreateCacheGroupContext$8(GridCacheProcessor.java:2415)
> at 
> org.apache.ignite.internal.util.InitializationProtector.protect(InitializationProtector.java:60)
> at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.getOrCreateCacheGroupContext(GridCacheProcessor.java:2412)
> at 
> 

[jira] [Created] (IGNITE-10504) If client have cache resource with not configurate data region it stop by handler

2018-12-03 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10504:
---

 Summary: If client have cache resource with not configurate data 
region it stop by handler
 Key: IGNITE-10504
 URL: https://issues.apache.org/jira/browse/IGNITE-10504
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10477) An empty cluster falls if wal segment size small

2018-11-29 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10477:

Description: 
I set   and try to activate 
empty cluster.
Get cluster drop by handler and next error
Two servers and many caches
{code:java}
[15:45:12,723][SEVERE][db-checkpoint-thread-#99][] Critical system error 
detected. Will be handled accordingly to configured handler [hnd=class 
o.a.i.failure.StopNodeOrHaltFailureHandler, failureCtx=FailureContext 
[type=SYSTEM_WORKER_TE
RMINATION, err=java.lang.IllegalArgumentException: Record is too long 
[capacity=100, size=1802204]]]
java.lang.IllegalArgumentException: Record is too long [capacity=100, 
size=1802204]
at 
org.apache.ignite.internal.processors.cache.persistence.wal.SegmentedRingByteBuffer.offer0(SegmentedRingByteBuffer.java:214)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.SegmentedRingByteBuffer.offer(SegmentedRingByteBuffer.java:193)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileWriteHandle.addRecord(FileWriteAheadLogManager.java:2472)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileWriteHandle.access$1600(FileWriteAheadLogManager.java:2376)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager.log(FileWriteAheadLogManager.java:821)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.markCheckpointBegin(GridCacheDatabaseSharedManager.java:3604)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.doCheckpoint(GridCacheDatabaseSharedManager.java:3091)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.body(GridCacheDatabaseSharedManager.java:2990)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
at java.lang.Thread.run(Thread.java:748)
{code}


  was:
I set   and try to activate 
empty cluster.
Get cluster drop by handler and next error
{code:java}
[15:45:12,723][SEVERE][db-checkpoint-thread-#99][] Critical system error 
detected. Will be handled accordingly to configured handler [hnd=class 
o.a.i.failure.StopNodeOrHaltFailureHandler, failureCtx=FailureContext 
[type=SYSTEM_WORKER_TE
RMINATION, err=java.lang.IllegalArgumentException: Record is too long 
[capacity=100, size=1802204]]]
java.lang.IllegalArgumentException: Record is too long [capacity=100, 
size=1802204]
at 
org.apache.ignite.internal.processors.cache.persistence.wal.SegmentedRingByteBuffer.offer0(SegmentedRingByteBuffer.java:214)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.SegmentedRingByteBuffer.offer(SegmentedRingByteBuffer.java:193)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileWriteHandle.addRecord(FileWriteAheadLogManager.java:2472)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileWriteHandle.access$1600(FileWriteAheadLogManager.java:2376)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager.log(FileWriteAheadLogManager.java:821)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.markCheckpointBegin(GridCacheDatabaseSharedManager.java:3604)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.doCheckpoint(GridCacheDatabaseSharedManager.java:3091)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.body(GridCacheDatabaseSharedManager.java:2990)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
at java.lang.Thread.run(Thread.java:748)
{code}



>  An empty cluster falls if wal segment size small
> -
>
> Key: IGNITE-10477
> URL: https://issues.apache.org/jira/browse/IGNITE-10477
> Project: Ignite
>  Issue Type: Bug
>  Components: data structures
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> I set   and try to activate 
> empty cluster.
> Get cluster drop by handler and next error
> Two servers and many caches
> {code:java}
> [15:45:12,723][SEVERE][db-checkpoint-thread-#99][] Critical system error 
> detected. Will be handled accordingly to configured handler [hnd=class 
> o.a.i.failure.StopNodeOrHaltFailureHandler, failureCtx=FailureContext 
> [type=SYSTEM_WORKER_TE
> RMINATION, err=java.lang.IllegalArgumentException: Record is too long 
> [capacity=100, size=1802204]]]

[jira] [Created] (IGNITE-10477) An empty cluster falls if wal segment size small

2018-11-29 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10477:
---

 Summary:  An empty cluster falls if wal segment size small
 Key: IGNITE-10477
 URL: https://issues.apache.org/jira/browse/IGNITE-10477
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8


I set   and try to activate 
empty cluster.
Get cluster drop by handler and next error
{code:java}
[15:45:12,723][SEVERE][db-checkpoint-thread-#99][] Critical system error 
detected. Will be handled accordingly to configured handler [hnd=class 
o.a.i.failure.StopNodeOrHaltFailureHandler, failureCtx=FailureContext 
[type=SYSTEM_WORKER_TE
RMINATION, err=java.lang.IllegalArgumentException: Record is too long 
[capacity=100, size=1802204]]]
java.lang.IllegalArgumentException: Record is too long [capacity=100, 
size=1802204]
at 
org.apache.ignite.internal.processors.cache.persistence.wal.SegmentedRingByteBuffer.offer0(SegmentedRingByteBuffer.java:214)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.SegmentedRingByteBuffer.offer(SegmentedRingByteBuffer.java:193)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileWriteHandle.addRecord(FileWriteAheadLogManager.java:2472)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager$FileWriteHandle.access$1600(FileWriteAheadLogManager.java:2376)
at 
org.apache.ignite.internal.processors.cache.persistence.wal.FileWriteAheadLogManager.log(FileWriteAheadLogManager.java:821)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.markCheckpointBegin(GridCacheDatabaseSharedManager.java:3604)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.doCheckpoint(GridCacheDatabaseSharedManager.java:3091)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheDatabaseSharedManager$Checkpointer.body(GridCacheDatabaseSharedManager.java:2990)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
at java.lang.Thread.run(Thread.java:748)
{code}




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10446) control.sh --cache idle_verify fail with NPE when node left grid

2018-11-28 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10446:

Description: 

Command [IDLE_VERIFY -host=***] started at [2018-11-28 15:54:23]...
Error code: 13000. java.lang.NullPointerException.
Command [IDLE_VERIFY] failed with error: 13000 - command failed.

> control.sh --cache idle_verify fail with NPE when node left grid
> 
>
> Key: IGNITE-10446
> URL: https://issues.apache.org/jira/browse/IGNITE-10446
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> 
> Command [IDLE_VERIFY -host=***] started at [2018-11-28 15:54:23]...
> Error code: 13000. java.lang.NullPointerException.
> Command [IDLE_VERIFY] failed with error: 13000 - command failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10446) control.sh --cache idle_verify fail with NPE when node left grid

2018-11-28 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10446:
---

 Summary: control.sh --cache idle_verify fail with NPE when node 
left grid
 Key: IGNITE-10446
 URL: https://issues.apache.org/jira/browse/IGNITE-10446
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10402) Restart cluster with BLT led to loss of node due to AE

2018-11-26 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10402:

Description: 
{code:java}
[23:44:47,616][SEVERE][sys-#88][] Critical system error detected. Will be 
handled accordingly to configured handler [hnd=StopNodeOrHaltFailureHandler 
[tryStop=false, timeout=0, super=AbstractFailureHandler 
[ignoredFailureTypes=Singleton
Set [SYSTEM_WORKER_BLOCKED]]], failureCtx=FailureContext [type=CRITICAL_ERROR, 
err=java.lang.AssertionError]]
java.lang.AssertionError
{code}


  was:

{code:java}
[23:44:47,616][SEVERE][sys-#88][] Critical system error detected. Will be 
handled accordingly to configured handler [hnd=StopNodeOrHaltFailureHandler 
[tryStop=false, timeout=0, super=AbstractFailureHandler 
[ignoredFailureTypes=Singleton
Set [SYSTEM_WORKER_BLOCKED]]], failureCtx=FailureContext [type=CRITICAL_ERROR, 
err=java.lang.AssertionError]]
java.lang.AssertionError
at 
org.gridgain.grid.internal.processors.cache.database.GridSnapshotAwareClusterStateProcessorImpl$3.apply(GridSnapshotAwareClusterStateProcessorImpl.java:291)
at 
org.gridgain.grid.internal.processors.cache.database.GridSnapshotAwareClusterStateProcessorImpl$3.apply(GridSnapshotAwareClusterStateProcessorImpl.java:270)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.notifyListener(GridFutureAdapter.java:385)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.unblock(GridFutureAdapter.java:349)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.unblockAll(GridFutureAdapter.java:337)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:497)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:476)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.onDone(GridDhtPartitionsExchangeFuture.java:2183)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.finishExchangeOnCoordinator(GridDhtPartitionsExchangeFuture.java:3497)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.onAllReceived(GridDhtPartitionsExchangeFuture.java:3208)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.processSingleMessage(GridDhtPartitionsExchangeFuture.java:2809)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.access$100(GridDhtPartitionsExchangeFuture.java:141)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture$2.apply(GridDhtPartitionsExchangeFuture.java:2614)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture$2.apply(GridDhtPartitionsExchangeFuture.java:2602)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.notifyListener(GridFutureAdapter.java:385)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.listen(GridFutureAdapter.java:355)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.onReceiveSingleMessage(GridDhtPartitionsExchangeFuture.java:2602)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager.processSinglePartitionUpdate(GridCachePartitionExchangeManager.java:1839)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager.access$1200(GridCachePartitionExchangeManager.java:146)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$2.onMessage(GridCachePartitionExchangeManager.java:386)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$2.onMessage(GridCachePartitionExchangeManager.java:344)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$MessageHandler.apply(GridCachePartitionExchangeManager.java:3334)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$MessageHandler.apply(GridCachePartitionExchangeManager.java:3313)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.processMessage(GridCacheIoManager.java:1061)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.onMessage0(GridCacheIoManager.java:586)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.handleMessage(GridCacheIoManager.java:385)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.handleMessage(GridCacheIoManager.java:311)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.access$100(GridCacheIoManager.java:101)
at 

[jira] [Resolved] (IGNITE-10402) Restart cluster with BLT led to loss of node due to AE

2018-11-26 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov resolved IGNITE-10402.
-
   Resolution: Cannot Reproduce
Fix Version/s: (was: 2.8)

> Restart cluster with BLT led to loss of node due to AE
> --
>
> Key: IGNITE-10402
> URL: https://issues.apache.org/jira/browse/IGNITE-10402
> Project: Ignite
>  Issue Type: Bug
>  Components: data structures
>Affects Versions: 2.8
>Reporter: ARomantsov
>Assignee: Sergey Antonov
>Priority: Critical
>
> {code:java}
> [23:44:47,616][SEVERE][sys-#88][] Critical system error detected. Will be 
> handled accordingly to configured handler [hnd=StopNodeOrHaltFailureHandler 
> [tryStop=false, timeout=0, super=AbstractFailureHandler 
> [ignoredFailureTypes=Singleton
> Set [SYSTEM_WORKER_BLOCKED]]], failureCtx=FailureContext 
> [type=CRITICAL_ERROR, err=java.lang.AssertionError]]
> java.lang.AssertionError
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10404) Handler stop node with reason - IgniteException: Checkpoint read lock acquisition has been timed out

2018-11-25 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10404:

Description: I know about possible WA  - set readLockCheckpointTimeout to 
zero, but it make product unstable, sometimes I forget this option and my 
cluster totally broken when I try to load  (was: I now about possible WA  - set 
readLockCheckpointTimeout to zero, but it make product unstable, sometimes I 
forget this option and my cluster totally broken when I try to load)

> Handler stop node with reason - IgniteException: Checkpoint read lock 
> acquisition has been timed out
> 
>
> Key: IGNITE-10404
> URL: https://issues.apache.org/jira/browse/IGNITE-10404
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> I know about possible WA  - set readLockCheckpointTimeout to zero, but it 
> make product unstable, sometimes I forget this option and my cluster totally 
> broken when I try to load



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10404) Handler stop node with reason - IgniteException: Checkpoint read lock acquisition has been timed out

2018-11-25 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10404:

Affects Version/s: 2.7

> Handler stop node with reason - IgniteException: Checkpoint read lock 
> acquisition has been timed out
> 
>
> Key: IGNITE-10404
> URL: https://issues.apache.org/jira/browse/IGNITE-10404
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
>
> I now about possible WA  - set readLockCheckpointTimeout to zero, but it make 
> product unstable, sometimes I forget this option and my cluster totally 
> broken when I try to load



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10404) Handler stop node with reason - IgniteException: Checkpoint read lock acquisition has been timed out

2018-11-25 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10404:

Fix Version/s: 2.8

> Handler stop node with reason - IgniteException: Checkpoint read lock 
> acquisition has been timed out
> 
>
> Key: IGNITE-10404
> URL: https://issues.apache.org/jira/browse/IGNITE-10404
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> I now about possible WA  - set readLockCheckpointTimeout to zero, but it make 
> product unstable, sometimes I forget this option and my cluster totally 
> broken when I try to load



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10404) Handler stop node with reason - IgniteException: Checkpoint read lock acquisition has been timed out

2018-11-25 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10404:
---

 Summary: Handler stop node with reason - IgniteException: 
Checkpoint read lock acquisition has been timed out
 Key: IGNITE-10404
 URL: https://issues.apache.org/jira/browse/IGNITE-10404
 Project: Ignite
  Issue Type: Bug
Reporter: ARomantsov


I now about possible WA  - set readLockCheckpointTimeout to zero, but it make 
product unstable, sometimes I forget this option and my cluster totally broken 
when I try to load



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10402) Restart cluster with BLT led to loss of node due to AE

2018-11-25 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10402:
---

 Summary: Restart cluster with BLT led to loss of node due to AE
 Key: IGNITE-10402
 URL: https://issues.apache.org/jira/browse/IGNITE-10402
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.8
Reporter: ARomantsov
 Fix For: 2.8



{code:java}
[23:44:47,616][SEVERE][sys-#88][] Critical system error detected. Will be 
handled accordingly to configured handler [hnd=StopNodeOrHaltFailureHandler 
[tryStop=false, timeout=0, super=AbstractFailureHandler 
[ignoredFailureTypes=Singleton
Set [SYSTEM_WORKER_BLOCKED]]], failureCtx=FailureContext [type=CRITICAL_ERROR, 
err=java.lang.AssertionError]]
java.lang.AssertionError
at 
org.gridgain.grid.internal.processors.cache.database.GridSnapshotAwareClusterStateProcessorImpl$3.apply(GridSnapshotAwareClusterStateProcessorImpl.java:291)
at 
org.gridgain.grid.internal.processors.cache.database.GridSnapshotAwareClusterStateProcessorImpl$3.apply(GridSnapshotAwareClusterStateProcessorImpl.java:270)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.notifyListener(GridFutureAdapter.java:385)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.unblock(GridFutureAdapter.java:349)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.unblockAll(GridFutureAdapter.java:337)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:497)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:476)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.onDone(GridDhtPartitionsExchangeFuture.java:2183)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.finishExchangeOnCoordinator(GridDhtPartitionsExchangeFuture.java:3497)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.onAllReceived(GridDhtPartitionsExchangeFuture.java:3208)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.processSingleMessage(GridDhtPartitionsExchangeFuture.java:2809)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.access$100(GridDhtPartitionsExchangeFuture.java:141)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture$2.apply(GridDhtPartitionsExchangeFuture.java:2614)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture$2.apply(GridDhtPartitionsExchangeFuture.java:2602)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.notifyListener(GridFutureAdapter.java:385)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.listen(GridFutureAdapter.java:355)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.preloader.GridDhtPartitionsExchangeFuture.onReceiveSingleMessage(GridDhtPartitionsExchangeFuture.java:2602)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager.processSinglePartitionUpdate(GridCachePartitionExchangeManager.java:1839)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager.access$1200(GridCachePartitionExchangeManager.java:146)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$2.onMessage(GridCachePartitionExchangeManager.java:386)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$2.onMessage(GridCachePartitionExchangeManager.java:344)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$MessageHandler.apply(GridCachePartitionExchangeManager.java:3334)
at 
org.apache.ignite.internal.processors.cache.GridCachePartitionExchangeManager$MessageHandler.apply(GridCachePartitionExchangeManager.java:3313)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.processMessage(GridCacheIoManager.java:1061)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.onMessage0(GridCacheIoManager.java:586)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.handleMessage(GridCacheIoManager.java:385)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.handleMessage(GridCacheIoManager.java:311)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager.access$100(GridCacheIoManager.java:101)
at 
org.apache.ignite.internal.processors.cache.GridCacheIoManager$1.onMessage(GridCacheIoManager.java:300)
at 

[jira] [Created] (IGNITE-10401) After node join and rebalance - some partition hang in move state

2018-11-25 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10401:
---

 Summary: After node join and rebalance - some partition hang in 
move state
 Key: IGNITE-10401
 URL: https://issues.apache.org/jira/browse/IGNITE-10401
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8


I try to kill three nodes and return them back after some load time

They begin to rebalance, but one host eventually have many MOVING partition and 
they not become owning after PME



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (IGNITE-10400) Server node stop by handler IgniteException: Failed to resolve node topology

2018-11-25 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov resolved IGNITE-10400.
-
Resolution: Duplicate

> Server node stop by handler IgniteException: Failed to resolve node topology
> 
>
> Key: IGNITE-10400
> URL: https://issues.apache.org/jira/browse/IGNITE-10400
> Project: Ignite
>  Issue Type: Bug
>Reporter: ARomantsov
>Priority: Critical
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10400) Server node stop by handler IgniteException: Failed to resolve node topology

2018-11-25 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10400:
---

 Summary: Server node stop by handler IgniteException: Failed to 
resolve node topology
 Key: IGNITE-10400
 URL: https://issues.apache.org/jira/browse/IGNITE-10400
 Project: Ignite
  Issue Type: Bug
Reporter: ARomantsov






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10395) Add to control.sh --cache --tx overall info: count of transaction and count of transaction in each state (Rollback, Commiting and etc)

2018-11-23 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10395:

Summary: Add to control.sh --cache --tx overall info: count of transaction 
and count of transaction in each state (Rollback, Commiting and etc)  (was: Add 
to control.sh --cache --tx overall info:)

> Add to control.sh --cache --tx overall info: count of transaction and count 
> of transaction in each state (Rollback, Commiting and etc)
> --
>
> Key: IGNITE-10395
> URL: https://issues.apache.org/jira/browse/IGNITE-10395
> Project: Ignite
>  Issue Type: Improvement
>Reporter: ARomantsov
>Priority: Major
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10395) Add to control.sh --cache --tx overall info:

2018-11-23 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10395:

Issue Type: Improvement  (was: Bug)

> Add to control.sh --cache --tx overall info:
> 
>
> Key: IGNITE-10395
> URL: https://issues.apache.org/jira/browse/IGNITE-10395
> Project: Ignite
>  Issue Type: Improvement
>Reporter: ARomantsov
>Priority: Major
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10395) Add to control.sh --cache --tx overall info:

2018-11-23 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10395:
---

 Summary: Add to control.sh --cache --tx overall info:
 Key: IGNITE-10395
 URL: https://issues.apache.org/jira/browse/IGNITE-10395
 Project: Ignite
  Issue Type: Bug
Reporter: ARomantsov






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10394) Try to activate cluster after deactivate. All node exit by handler

2018-11-23 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10394:
---

 Summary: Try to activate cluster after deactivate. All node exit 
by handler
 Key: IGNITE-10394
 URL: https://issues.apache.org/jira/browse/IGNITE-10394
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.7
Reporter: ARomantsov


AE: ignite-sys-cache
..processors.cache.CacheRegistry.update(CacheRegistry.java:188)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10205) add to utility command - ./control.sh --cache idle_verify --dump abbility to exclude cache from output file

2018-11-20 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10205:

Priority: Major  (was: Critical)

> add to utility command -  ./control.sh --cache idle_verify --dump abbility to 
> exclude cache from output file
> 
>
> Key: IGNITE-10205
> URL: https://issues.apache.org/jira/browse/IGNITE-10205
> Project: Ignite
>  Issue Type: Improvement
>Affects Versions: 2.6
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10327) Claster take NPE due cache stop and stop by handler

2018-11-19 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10327:
---

 Summary: Claster take NPE due cache stop and stop by handler
 Key: IGNITE-10327
 URL: https://issues.apache.org/jira/browse/IGNITE-10327
 Project: Ignite
  Issue Type: Bug
  Components: data structures
Affects Versions: 2.8
Reporter: ARomantsov
 Fix For: 2.7


Stop all caches
Take NPE : null on preloadEntry(GridDhtPartitionDemander.java:893)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10323) Contol utility --deactivate on non-activate cluster produse NPE and handler stop nodes

2018-11-19 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10323:
---

 Summary: Contol utility --deactivate on non-activate cluster 
produse NPE and handler stop nodes
 Key: IGNITE-10323
 URL: https://issues.apache.org/jira/browse/IGNITE-10323
 Project: Ignite
  Issue Type: Bug
  Components: persistence
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8


Scenario:
1) Start cluster
2) Call control.sh --deactivate




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10313) Long exchange on deactivation process

2018-11-18 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10313:
---

 Summary: Long exchange on deactivation process
 Key: IGNITE-10313
 URL: https://issues.apache.org/jira/browse/IGNITE-10313
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.7
 Environment: 16 host, 1 server node per host
Reporter: ARomantsov
 Fix For: 2.8


Long exhange after start deactivation process - near to three minutes.
Probably in doesn't end, but control.sh return to control to console and I stop 
cluster



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9980) Modify ./control.sh --cache idle_verify --dump print to diff mode (user persistant only/user not-persistent only/system only) cache

2018-11-09 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9980?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9980:
---
Fix Version/s: 2.8

> Modify ./control.sh --cache idle_verify --dump print to diff mode (user 
> persistant only/user not-persistent only/system only) cache
> ---
>
> Key: IGNITE-9980
> URL: https://issues.apache.org/jira/browse/IGNITE-9980
> Project: Ignite
>  Issue Type: Improvement
>Affects Versions: 2.6
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> It will be cool , if control.sh --cache idle_verify can show 
> persistent/not-persistent/system caches and it will be impliments due utility 
> flag



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10205) add to utility command - ./control.sh --cache idle_verify --dump abbility to exclude cache from output file

2018-11-09 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10205:

Affects Version/s: 2.6

> add to utility command -  ./control.sh --cache idle_verify --dump abbility to 
> exclude cache from output file
> 
>
> Key: IGNITE-10205
> URL: https://issues.apache.org/jira/browse/IGNITE-10205
> Project: Ignite
>  Issue Type: Improvement
>Affects Versions: 2.6
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-10205) add to utility command - ./control.sh --cache idle_verify --dump abbility to exclude cache from output file

2018-11-09 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10205:

Fix Version/s: 2.8

> add to utility command -  ./control.sh --cache idle_verify --dump abbility to 
> exclude cache from output file
> 
>
> Key: IGNITE-10205
> URL: https://issues.apache.org/jira/browse/IGNITE-10205
> Project: Ignite
>  Issue Type: Improvement
>Affects Versions: 2.6
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-10205) add to utility command - ./control.sh --cache idle_verify --dump abbility to exclude cache from output file

2018-11-09 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10205:
---

 Summary: add to utility command -  ./control.sh --cache 
idle_verify --dump abbility to exclude cache from output file
 Key: IGNITE-10205
 URL: https://issues.apache.org/jira/browse/IGNITE-10205
 Project: Ignite
  Issue Type: Improvement
Reporter: ARomantsov






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9980) Modify ./control.sh --cache idle_verify --dump print to diff mode (user persistant only/user not-persistent only/system only) cache

2018-11-09 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9980?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9980:
---
Issue Type: Improvement  (was: Bug)

> Modify ./control.sh --cache idle_verify --dump print to diff mode (user 
> persistant only/user not-persistent only/system only) cache
> ---
>
> Key: IGNITE-9980
> URL: https://issues.apache.org/jira/browse/IGNITE-9980
> Project: Ignite
>  Issue Type: Improvement
>Affects Versions: 2.6
>Reporter: ARomantsov
>Priority: Major
> Fix For: 2.8
>
>
> It will be cool , if control.sh --cache idle_verify can show 
> persistent/not-persistent/system caches and it will be impliments due utility 
> flag



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9980) Modify ./control.sh --cache idle_verify --dump print to diff mode (user persistant only/user not-persistent only/system only) cache

2018-11-09 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9980?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9980:
---
Description: It will be cool , if control.sh --cache idle_verify can show 
persistent/not-persistent/system caches and it will be impliments due utility 
flag  (was: It will be cool , if control.sh --cache idle_verify can show 
persistent/not-persistent/both caches and it will be impliments due utility 
flag)

> Modify ./control.sh --cache idle_verify --dump print to diff mode (user 
> persistant only/user not-persistent only/system only) cache
> ---
>
> Key: IGNITE-9980
> URL: https://issues.apache.org/jira/browse/IGNITE-9980
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.6
>Reporter: ARomantsov
>Priority: Major
>
> It will be cool , if control.sh --cache idle_verify can show 
> persistent/not-persistent/system caches and it will be impliments due utility 
> flag



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9980) Modify ./control.sh --cache idle_verify --dump print to diff mode (user persistant only/user not-persistent only/system only) cache

2018-11-09 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9980?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9980:
---
Summary: Modify ./control.sh --cache idle_verify --dump print to diff mode 
(user persistant only/user not-persistent only/system only) cache  (was: Modify 
./control.sh --cache idle_verify --dump to print only persistante cache or add 
flag --persistentOnly)

> Modify ./control.sh --cache idle_verify --dump print to diff mode (user 
> persistant only/user not-persistent only/system only) cache
> ---
>
> Key: IGNITE-9980
> URL: https://issues.apache.org/jira/browse/IGNITE-9980
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.6
>Reporter: ARomantsov
>Priority: Major
>
> It will be cool , if control.sh --cache idle_verify can show 
> persistent/not-persistent/both caches and it will be impliments due utility 
> flag



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (IGNITE-10134) control.sh --cache list . --config not working on big cache setup

2018-11-02 Thread ARomantsov (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-10134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16673299#comment-16673299
 ] 

ARomantsov commented on IGNITE-10134:
-

I found that command on - https://issues.apache.org/jira/browse/IGNITE-9853

> control.sh --cache list . --config not working on big cache setup
> -
>
> Key: IGNITE-10134
> URL: https://issues.apache.org/jira/browse/IGNITE-10134
> Project: Ignite
>  Issue Type: Bug
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> {code:java}
> [17:09:31,352][SEVERE][mgmt-#271][GridTaskWorker] Failed to reduce job 
> results due to undeclared user exception 
> [task=o.a.i.i.v.cache.VisorCacheConfigurationCollectorTask@22c7b357, 
> err=java.lang.AssertionError]
> java.lang.AssertionError
> at 
> org.apache.ignite.internal.visor.VisorOneNodeTask.reduce0(VisorOneNodeTask.java:30)
> at 
> org.apache.ignite.internal.visor.VisorMultiNodeTask.reduce(VisorMultiNodeTask.java:139)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker$6.call(GridTaskWorker.java:1133)
> at 
> org.apache.ignite.internal.util.IgniteUtils.wrapThreadLoader(IgniteUtils.java:6726)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.reduce(GridTaskWorker.java:1131)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.onResponse(GridTaskWorker.java:958)
> at 
> org.apache.ignite.internal.processors.task.GridTaskProcessor.processJobExecuteResponse(GridTaskProcessor.java:1104)
> at 
> org.apache.ignite.internal.processors.task.GridTaskProcessor$JobMessageListener.onMessage(GridTaskProcessor.java:1339)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager.invokeListener(GridIoManager.java:1556)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager.processRegularMessage0(GridIoManager.java:1184)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager.access$4200(GridIoManager.java:125)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager$9.run(GridIoManager.java:1091)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> [17:09:31,354][SEVERE][mgmt-#271][GridTaskCommandHandler] Failed to execute 
> task [name=o.a.i.i.v.cache.VisorCacheConfigurationCollectorTask, 
> clientId=606bf232-20e0-4799-bfb1-8d6e44b0d1ad]
> class org.apache.ignite.IgniteCheckedException: Failed to reduce job results 
> due to undeclared user exception 
> [task=org.apache.ignite.internal.visor.cache.VisorCacheConfigurationCollectorTask@22c7b357,
>  err=java.lang.AssertionError]
> at 
> org.apache.ignite.internal.util.IgniteUtils.cast(IgniteUtils.java:7338)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.resolve(GridFutureAdapter.java:259)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.get0(GridFutureAdapter.java:171)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.get(GridFutureAdapter.java:140)
> at 
> org.apache.ignite.internal.processors.rest.handlers.task.GridTaskCommandHandler$2.apply(GridTaskCommandHandler.java:263)
> at 
> org.apache.ignite.internal.processors.rest.handlers.task.GridTaskCommandHandler$2.apply(GridTaskCommandHandler.java:257)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.notifyListener(GridFutureAdapter.java:383)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.unblock(GridFutureAdapter.java:347)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.unblockAll(GridFutureAdapter.java:335)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:495)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:474)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.finishTask(GridTaskWorker.java:1617)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.finishTask(GridTaskWorker.java:1585)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.reduce(GridTaskWorker.java:1175)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.onResponse(GridTaskWorker.java:958)
> at 
> org.apache.ignite.internal.processors.task.GridTaskProcessor.processJobExecuteResponse(GridTaskProcessor.java:1104)
> at 
> org.apache.ignite.internal.processors.task.GridTaskProcessor$JobMessageListener.onMessage(GridTaskProcessor.java:1339)
> at 
> 

[jira] [Created] (IGNITE-10134) control.sh --cache list . --config not working on big cache setup

2018-11-02 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10134:
---

 Summary: control.sh --cache list . --config not working on big 
cache setup
 Key: IGNITE-10134
 URL: https://issues.apache.org/jira/browse/IGNITE-10134
 Project: Ignite
  Issue Type: Bug
Reporter: ARomantsov



{code:java}
[17:09:31,352][SEVERE][mgmt-#271][GridTaskWorker] Failed to reduce job results 
due to undeclared user exception 
[task=o.a.i.i.v.cache.VisorCacheConfigurationCollectorTask@22c7b357, 
err=java.lang.AssertionError]
java.lang.AssertionError
at 
org.apache.ignite.internal.visor.VisorOneNodeTask.reduce0(VisorOneNodeTask.java:30)
at 
org.apache.ignite.internal.visor.VisorMultiNodeTask.reduce(VisorMultiNodeTask.java:139)
at 
org.apache.ignite.internal.processors.task.GridTaskWorker$6.call(GridTaskWorker.java:1133)
at 
org.apache.ignite.internal.util.IgniteUtils.wrapThreadLoader(IgniteUtils.java:6726)
at 
org.apache.ignite.internal.processors.task.GridTaskWorker.reduce(GridTaskWorker.java:1131)
at 
org.apache.ignite.internal.processors.task.GridTaskWorker.onResponse(GridTaskWorker.java:958)
at 
org.apache.ignite.internal.processors.task.GridTaskProcessor.processJobExecuteResponse(GridTaskProcessor.java:1104)
at 
org.apache.ignite.internal.processors.task.GridTaskProcessor$JobMessageListener.onMessage(GridTaskProcessor.java:1339)
at 
org.apache.ignite.internal.managers.communication.GridIoManager.invokeListener(GridIoManager.java:1556)
at 
org.apache.ignite.internal.managers.communication.GridIoManager.processRegularMessage0(GridIoManager.java:1184)
at 
org.apache.ignite.internal.managers.communication.GridIoManager.access$4200(GridIoManager.java:125)
at 
org.apache.ignite.internal.managers.communication.GridIoManager$9.run(GridIoManager.java:1091)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
[17:09:31,354][SEVERE][mgmt-#271][GridTaskCommandHandler] Failed to execute 
task [name=o.a.i.i.v.cache.VisorCacheConfigurationCollectorTask, 
clientId=606bf232-20e0-4799-bfb1-8d6e44b0d1ad]
class org.apache.ignite.IgniteCheckedException: Failed to reduce job results 
due to undeclared user exception 
[task=org.apache.ignite.internal.visor.cache.VisorCacheConfigurationCollectorTask@22c7b357,
 err=java.lang.AssertionError]
at 
org.apache.ignite.internal.util.IgniteUtils.cast(IgniteUtils.java:7338)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.resolve(GridFutureAdapter.java:259)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.get0(GridFutureAdapter.java:171)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.get(GridFutureAdapter.java:140)
at 
org.apache.ignite.internal.processors.rest.handlers.task.GridTaskCommandHandler$2.apply(GridTaskCommandHandler.java:263)
at 
org.apache.ignite.internal.processors.rest.handlers.task.GridTaskCommandHandler$2.apply(GridTaskCommandHandler.java:257)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.notifyListener(GridFutureAdapter.java:383)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.unblock(GridFutureAdapter.java:347)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.unblockAll(GridFutureAdapter.java:335)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:495)
at 
org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:474)
at 
org.apache.ignite.internal.processors.task.GridTaskWorker.finishTask(GridTaskWorker.java:1617)
at 
org.apache.ignite.internal.processors.task.GridTaskWorker.finishTask(GridTaskWorker.java:1585)
at 
org.apache.ignite.internal.processors.task.GridTaskWorker.reduce(GridTaskWorker.java:1175)
at 
org.apache.ignite.internal.processors.task.GridTaskWorker.onResponse(GridTaskWorker.java:958)
at 
org.apache.ignite.internal.processors.task.GridTaskProcessor.processJobExecuteResponse(GridTaskProcessor.java:1104)
at 
org.apache.ignite.internal.processors.task.GridTaskProcessor$JobMessageListener.onMessage(GridTaskProcessor.java:1339)
at 
org.apache.ignite.internal.managers.communication.GridIoManager.invokeListener(GridIoManager.java:1556)
at 
org.apache.ignite.internal.managers.communication.GridIoManager.processRegularMessage0(GridIoManager.java:1184)
at 
org.apache.ignite.internal.managers.communication.GridIoManager.access$4200(GridIoManager.java:125)
at 
org.apache.ignite.internal.managers.communication.GridIoManager$9.run(GridIoManager.java:1091)
at 

[jira] [Updated] (IGNITE-10134) control.sh --cache list . --config not working on big cache setup

2018-11-02 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-10134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-10134:

Fix Version/s: 2.8

> control.sh --cache list . --config not working on big cache setup
> -
>
> Key: IGNITE-10134
> URL: https://issues.apache.org/jira/browse/IGNITE-10134
> Project: Ignite
>  Issue Type: Bug
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.8
>
>
> {code:java}
> [17:09:31,352][SEVERE][mgmt-#271][GridTaskWorker] Failed to reduce job 
> results due to undeclared user exception 
> [task=o.a.i.i.v.cache.VisorCacheConfigurationCollectorTask@22c7b357, 
> err=java.lang.AssertionError]
> java.lang.AssertionError
> at 
> org.apache.ignite.internal.visor.VisorOneNodeTask.reduce0(VisorOneNodeTask.java:30)
> at 
> org.apache.ignite.internal.visor.VisorMultiNodeTask.reduce(VisorMultiNodeTask.java:139)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker$6.call(GridTaskWorker.java:1133)
> at 
> org.apache.ignite.internal.util.IgniteUtils.wrapThreadLoader(IgniteUtils.java:6726)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.reduce(GridTaskWorker.java:1131)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.onResponse(GridTaskWorker.java:958)
> at 
> org.apache.ignite.internal.processors.task.GridTaskProcessor.processJobExecuteResponse(GridTaskProcessor.java:1104)
> at 
> org.apache.ignite.internal.processors.task.GridTaskProcessor$JobMessageListener.onMessage(GridTaskProcessor.java:1339)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager.invokeListener(GridIoManager.java:1556)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager.processRegularMessage0(GridIoManager.java:1184)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager.access$4200(GridIoManager.java:125)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager$9.run(GridIoManager.java:1091)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> [17:09:31,354][SEVERE][mgmt-#271][GridTaskCommandHandler] Failed to execute 
> task [name=o.a.i.i.v.cache.VisorCacheConfigurationCollectorTask, 
> clientId=606bf232-20e0-4799-bfb1-8d6e44b0d1ad]
> class org.apache.ignite.IgniteCheckedException: Failed to reduce job results 
> due to undeclared user exception 
> [task=org.apache.ignite.internal.visor.cache.VisorCacheConfigurationCollectorTask@22c7b357,
>  err=java.lang.AssertionError]
> at 
> org.apache.ignite.internal.util.IgniteUtils.cast(IgniteUtils.java:7338)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.resolve(GridFutureAdapter.java:259)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.get0(GridFutureAdapter.java:171)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.get(GridFutureAdapter.java:140)
> at 
> org.apache.ignite.internal.processors.rest.handlers.task.GridTaskCommandHandler$2.apply(GridTaskCommandHandler.java:263)
> at 
> org.apache.ignite.internal.processors.rest.handlers.task.GridTaskCommandHandler$2.apply(GridTaskCommandHandler.java:257)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.notifyListener(GridFutureAdapter.java:383)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.unblock(GridFutureAdapter.java:347)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.unblockAll(GridFutureAdapter.java:335)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:495)
> at 
> org.apache.ignite.internal.util.future.GridFutureAdapter.onDone(GridFutureAdapter.java:474)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.finishTask(GridTaskWorker.java:1617)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.finishTask(GridTaskWorker.java:1585)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.reduce(GridTaskWorker.java:1175)
> at 
> org.apache.ignite.internal.processors.task.GridTaskWorker.onResponse(GridTaskWorker.java:958)
> at 
> org.apache.ignite.internal.processors.task.GridTaskProcessor.processJobExecuteResponse(GridTaskProcessor.java:1104)
> at 
> org.apache.ignite.internal.processors.task.GridTaskProcessor$JobMessageListener.onMessage(GridTaskProcessor.java:1339)
> at 
> org.apache.ignite.internal.managers.communication.GridIoManager.invokeListener(GridIoManager.java:1556)

[jira] [Created] (IGNITE-10065) possibility to get EVT_CACHE_STARTED before cache proxy initialization

2018-10-30 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-10065:
---

 Summary: possibility to get EVT_CACHE_STARTED before cache proxy 
initialization
 Key: IGNITE-10065
 URL: https://issues.apache.org/jira/browse/IGNITE-10065
 Project: Ignite
  Issue Type: Bug
  Components: cache
Affects Versions: 2.7
Reporter: ARomantsov
 Fix For: 2.8


It looks , I can take EVT_CACHE_STARTED before take message -  Finish proxy 
initialization, cacheName=cache



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9994) idle_verify -analyze -partId= -grpId=(minus)number not working

2018-10-25 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9994?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9994:
---
Priority: Major  (was: Critical)

> idle_verify -analyze -partId= -grpId=(minus)number not working
> --
>
> Key: IGNITE-9994
> URL: https://issues.apache.org/jira/browse/IGNITE-9994
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.6
>Reporter: ARomantsov
>Priority: Major
>
> If enter command with plus -grpId - it working



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-9994) idle_verify -analyze -partId= -grpId=(minus)number not working

2018-10-25 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-9994:
--

 Summary: idle_verify -analyze -partId= -grpId=(minus)number not 
working
 Key: IGNITE-9994
 URL: https://issues.apache.org/jira/browse/IGNITE-9994
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.6
Reporter: ARomantsov


If enter command with plus -grpId - it working



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9992) Add some command to calculate hast sum per primary partition in product

2018-10-24 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9992?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9992:
---
Description: 
Some util to quick check cluster data is ok
Or mode to control.sh --cache idle_verify --dump to calculate primary without 
log node info

Example
Cache - 1:
Part - hash
1 - 13123213213
2 - 1313123213
... etc


  was:
Some util to quick check cluster data is ok
Or mode to control.sh --cache idle_verify --dump to calculate primary without 
log node info


> Add some command to calculate hast sum per primary partition in product
> ---
>
> Key: IGNITE-9992
> URL: https://issues.apache.org/jira/browse/IGNITE-9992
> Project: Ignite
>  Issue Type: Improvement
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Major
>
> Some util to quick check cluster data is ok
> Or mode to control.sh --cache idle_verify --dump to calculate primary without 
> log node info
> Example
> Cache - 1:
> Part - hash
> 1 - 13123213213
> 2 - 1313123213
> ... etc



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9992) Add some command to calculate hast sum per primary partition in product

2018-10-24 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9992?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9992:
---
Description: 
Some util to quick check cluster data is ok
Or mode to control.sh --cache idle_verify --dump to calculate primary without 
log node info

  was:Some util to quick check cluster data is ok


> Add some command to calculate hast sum per primary partition in product
> ---
>
> Key: IGNITE-9992
> URL: https://issues.apache.org/jira/browse/IGNITE-9992
> Project: Ignite
>  Issue Type: Improvement
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Major
>
> Some util to quick check cluster data is ok
> Or mode to control.sh --cache idle_verify --dump to calculate primary without 
> log node info



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-9992) Add some command to calculate hast sum per primary partition in product

2018-10-24 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-9992:
--

 Summary: Add some command to calculate hast sum per primary 
partition in product
 Key: IGNITE-9992
 URL: https://issues.apache.org/jira/browse/IGNITE-9992
 Project: Ignite
  Issue Type: Improvement
Affects Versions: 2.7
Reporter: ARomantsov


Some util to quick check cluster data is ok



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (IGNITE-9980) Modify ./control.sh --cache idle_verify --dump to print only persistante cache or add flag --persistentOnly

2018-10-23 Thread ARomantsov (JIRA)
ARomantsov created IGNITE-9980:
--

 Summary: Modify ./control.sh --cache idle_verify --dump to print 
only persistante cache or add flag --persistentOnly
 Key: IGNITE-9980
 URL: https://issues.apache.org/jira/browse/IGNITE-9980
 Project: Ignite
  Issue Type: Bug
Affects Versions: 2.6
Reporter: ARomantsov


It will be cool , if control.sh --cache idle_verify can show 
persistent/not-persistent/both caches and it will be impliments due utility flag



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9905) After transaction load cluster inconsistent

2018-10-17 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9905:
---
Fix Version/s: (was: 2.7)

> After transaction load cluster inconsistent
> ---
>
> Key: IGNITE-9905
> URL: https://issues.apache.org/jira/browse/IGNITE-9905
> Project: Ignite
>  Issue Type: Bug
>  Components: persistence
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
>
> Loaded data into the cluster using transactions consisting of two get / two 
> put
> Test env: one server, two server node, one client
> {code:java}
> idle_verify check has finished, found 60 conflict partitions: 
> [counterConflicts=45, hashConflicts=15]
> Update counter conflicts:
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=98]
> Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
> consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
> PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
> size=596, partHash=-1167688484]]
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=34]
> Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
> consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
> PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
> size=596, partHash=-1284437377]]
> Conflict partition: PartitionKeyV2 [grpId=770187303, 
> grpName=CACHEGROUP_PARTICLE_1, partId=31]
> Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
> consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
> PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
> size=4, partHash=-1125172674]]
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=39]
> Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
> consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
> PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
> size=596, partHash=-40303136]]
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=90]
> Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
> consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
> PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
> size=596, partHash=-1221175703]]
> ...
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-9905) After transaction load cluster inconsistent

2018-10-16 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9905:
---
Description: 
Loaded data into the cluster using transactions consisting of two get / two put
Test env: one server, two server node, one client

{code:java}
idle_verify check has finished, found 60 conflict partitions: 
[counterConflicts=45, hashConflicts=15]
Update counter conflicts:
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=98]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
size=596, partHash=-1167688484]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=34]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
size=596, partHash=-1284437377]]
Conflict partition: PartitionKeyV2 [grpId=770187303, 
grpName=CACHEGROUP_PARTICLE_1, partId=31]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
size=4, partHash=-1125172674]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=39]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
size=596, partHash=-40303136]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=90]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
size=596, partHash=-1221175703]]

...

{code}

  was:


{code:java}
idle_verify check has finished, found 60 conflict partitions: 
[counterConflicts=45, hashConflicts=15]
Update counter conflicts:
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=98]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
size=596, partHash=-1167688484]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=34]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
size=596, partHash=-1284437377]]
Conflict partition: PartitionKeyV2 [grpId=770187303, 
grpName=CACHEGROUP_PARTICLE_1, partId=31]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
size=4, partHash=-1125172674]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=39]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
size=596, partHash=-40303136]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=90]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
size=596, partHash=-1221175703]]

...

{code}


> After transaction load cluster inconsistent
> ---
>
> Key: IGNITE-9905
> URL: https://issues.apache.org/jira/browse/IGNITE-9905
> Project: Ignite
>  Issue Type: Bug
>  Components: persistence
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.7
>
>
> Loaded data into the cluster using transactions consisting of two get / two 
> put
> Test env: one server, two server node, one client
> {code:java}
> idle_verify check has finished, found 60 conflict partitions: 
> [counterConflicts=45, hashConflicts=15]
> Update counter conflicts:
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=98]
> Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
> consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
> 

[jira] [Updated] (IGNITE-9905) After transaction load cluster inconsistent

2018-10-16 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9905:
---
Description: 


{code:java}
idle_verify check has finished, found 60 conflict partitions: 
[counterConflicts=45, hashConflicts=15]
Update counter conflicts:
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=98]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
size=596, partHash=-1167688484]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=34]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
size=596, partHash=-1284437377]]
Conflict partition: PartitionKeyV2 [grpId=770187303, 
grpName=CACHEGROUP_PARTICLE_1, partId=31]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
size=4, partHash=-1125172674]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=39]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
size=596, partHash=-40303136]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=90]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
size=596, partHash=-1221175703]]

...

{code}

  was:
{code:java}
idle_verify check has finished, found 60 conflict partitions: 
[counterConflicts=45, hashConflicts=15]
Update counter conflicts:
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=98]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
size=596, partHash=-1167688484]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=34]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
size=596, partHash=-1284437377]]
Conflict partition: PartitionKeyV2 [grpId=770187303, 
grpName=CACHEGROUP_PARTICLE_1, partId=31]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
size=4, partHash=-1125172674]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=39]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
size=596, partHash=-40303136]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=90]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
size=596, partHash=-1221175703]]

...

{code}


> After transaction load cluster inconsistent
> ---
>
> Key: IGNITE-9905
> URL: https://issues.apache.org/jira/browse/IGNITE-9905
> Project: Ignite
>  Issue Type: Bug
>  Components: persistence
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.7
>
>
> {code:java}
> idle_verify check has finished, found 60 conflict partitions: 
> [counterConflicts=45, hashConflicts=15]
> Update counter conflicts:
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=98]
> Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
> consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
> PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
> size=596, partHash=-1167688484]]
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=34]
> Partition instances: [PartitionHashRecordV2 

[jira] [Updated] (IGNITE-9905) After transaction load cluster inconsistent

2018-10-16 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9905:
---
Description: 
{code:java}
idle_verify check has finished, found 60 conflict partitions: 
[counterConflicts=45, hashConflicts=15]
Update counter conflicts:
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=98]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
size=596, partHash=-1167688484]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=34]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
size=596, partHash=-1284437377]]
Conflict partition: PartitionKeyV2 [grpId=770187303, 
grpName=CACHEGROUP_PARTICLE_1, partId=31]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
size=4, partHash=-1125172674]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=39]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
size=596, partHash=-40303136]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=90]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
size=596, partHash=-1221175703]]

...

{code}

  was:
{code:java}
idle_verify check has finished, found 60 conflict partitions: 
[counterConflicts=45, hashConflicts=15]
Update counter conflicts:
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=98]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
size=596, partHash=-1167688484]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=34]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
size=596, partHash=-1284437377]]
Conflict partition: PartitionKeyV2 [grpId=770187303, 
grpName=CACHEGROUP_PARTICLE_1, partId=31]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
size=4, partHash=-1125172674]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=39]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
size=596, partHash=-40303136]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=90]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
size=596, partHash=-1221175703]]



{code}


> After transaction load cluster inconsistent
> ---
>
> Key: IGNITE-9905
> URL: https://issues.apache.org/jira/browse/IGNITE-9905
> Project: Ignite
>  Issue Type: Bug
>  Components: persistence
>Affects Versions: 2.7
>Reporter: ARomantsov
>Priority: Critical
> Fix For: 2.7
>
>
> {code:java}
> idle_verify check has finished, found 60 conflict partitions: 
> [counterConflicts=45, hashConflicts=15]
> Update counter conflicts:
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=98]
> Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
> consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
> PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
> size=596, partHash=-1167688484]]
> Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
> grpName=CACHEGROUP_PARTICLE_1, partId=34]
> Partition instances: [PartitionHashRecordV2 

[jira] [Updated] (IGNITE-9905) After transaction load cluster inconsistent

2018-10-16 Thread ARomantsov (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-9905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ARomantsov updated IGNITE-9905:
---
Description: 
{code:java}
idle_verify check has finished, found 60 conflict partitions: 
[counterConflicts=45, hashConflicts=15]
Update counter conflicts:
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=98]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
size=596, partHash=-1167688484]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=34]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
size=596, partHash=-1284437377]]
Conflict partition: PartitionKeyV2 [grpId=770187303, 
grpName=CACHEGROUP_PARTICLE_1, partId=31]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
size=4, partHash=-1125172674]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=39]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
size=596, partHash=-40303136]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_1, partId=90]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
size=596, partHash=-1221175703]]



{code}

  was:
One server, two nodes per server, one client
Loaded data into the cluster using transactions consisting of two get / two put

{code:java}
idle_verify check has finished, found 60 conflict partitions: 
[counterConflicts=45, hashConflicts=15]
Update counter conflicts:
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_union-module_com.sbt.bm.ucp.common.dpl.model.dictionary.DServiceZone,
 partId=98]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1519, size=596, partHash=-1167688484], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1520, 
size=596, partHash=-1167688484]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_union-module_com.sbt.bm.ucp.common.dpl.model.dictionary.DServiceZone,
 partId=34]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1539, size=596, partHash=-99631005], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1537, 
size=596, partHash=-1284437377]]
Conflict partition: PartitionKeyV2 [grpId=770187303, 
grpName=CACHEGROUP_PARTICLE_currency-rates_com.sbt.cdm.api.model.dictionary.RKOKCostLevels,
 partId=31]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=15, size=4, partHash=-1125172674], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=16, 
size=4, partHash=-1125172674]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_union-module_com.sbt.bm.ucp.common.dpl.model.dictionary.DServiceZone,
 partId=39]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1555, size=596, partHash=-40303136], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1556, 
size=596, partHash=-40303136]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_union-module_com.sbt.bm.ucp.common.dpl.model.dictionary.DServiceZone,
 partId=90]
Partition instances: [PartitionHashRecordV2 [isPrimary=false, 
consistentId=node2, updateCntr=1557, size=596, partHash=-1295145299], 
PartitionHashRecordV2 [isPrimary=true, consistentId=node1, updateCntr=1556, 
size=596, partHash=-1221175703]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_union-module_com.sbt.bm.ucp.common.dpl.model.dictionary.DServiceZone,
 partId=28]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 
consistentId=node2, updateCntr=1480, size=596, partHash=840625505], 
PartitionHashRecordV2 [isPrimary=false, consistentId=node1, updateCntr=1481, 
size=596, partHash=840625505]]
Conflict partition: PartitionKeyV2 [grpId=-1903385190, 
grpName=CACHEGROUP_PARTICLE_union-module_com.sbt.bm.ucp.common.dpl.model.dictionary.DServiceZone,
 partId=29]
Partition instances: [PartitionHashRecordV2 [isPrimary=true, 

  1   2   >