[jira] [Resolved] (HDFS-15610) Reduce datanode upgrade/hardlink thread

2020-10-08 Thread Lokesh Jain (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDFS-15610.

Fix Version/s: 3.4.0
   Resolution: Fixed

> Reduce datanode upgrade/hardlink thread
> ---
>
> Key: HDFS-15610
> URL: https://issues.apache.org/jira/browse/HDFS-15610
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 3.0.0, 3.1.4
>Reporter: Karthik Palanisamy
>Assignee: Karthik Palanisamy
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> There is a kernel overhead on datanode upgrade. If datanode with millions of 
> blocks and 10+ disks then block-layout migration will be super expensive 
> during its hardlink operation.  Slowness is observed when running with large 
> hardlink threads(dfs.datanode.block.id.layout.upgrade.threads, default is 12 
> thread for each disk) and its runs for 2+ hours. 
> I.e 10*12=120 threads (for 10 disks)
> Small test:
> RHEL7, 32 cores, 20 GB RAM, 8 GB DN heap
> ||dfs.datanode.block.id.layout.upgrade.threads||Blocks||Disks||Time taken||
> |12|3.3 Million|1|2 minutes and 59 seconds|
> |6|3.3 Million|1|2 minutes and 35 seconds|
> |3|3.3 Million|1|2 minutes and 51 seconds|
> Tried same test twice and 95% is accurate (only a few sec difference on each 
> iteration). Using 6 thread is faster than 12 thread because of its overhead. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-15401) Namenode should log warning if concat/append finds file with large number of blocks

2020-06-09 Thread Lokesh Jain (Jira)
Lokesh Jain created HDFS-15401:
--

 Summary: Namenode should log warning if concat/append finds file 
with large number of blocks
 Key: HDFS-15401
 URL: https://issues.apache.org/jira/browse/HDFS-15401
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Lokesh Jain


Namenode should log warning if concat/append finds file has more than 
configured number of blocks. 

This is based on [~weichiu]'s comment 
https://issues.apache.org/jira/browse/HDFS-15392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17128732#comment-17128732.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-15400) fsck should log a warning if it finds a file with large number of blocks

2020-06-09 Thread Lokesh Jain (Jira)
Lokesh Jain created HDFS-15400:
--

 Summary: fsck should log a warning if it finds a file with large 
number of blocks
 Key: HDFS-15400
 URL: https://issues.apache.org/jira/browse/HDFS-15400
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Lokesh Jain


fsck should log a warning if it finds a file has more than configured number of 
blocks.

This is based on [~weichiu]'s comment 
https://issues.apache.org/jira/browse/HDFS-15392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17128732#comment-17128732.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-15392) DistrbutedFileSystem#concat api can create large number of small blocks

2020-06-05 Thread Lokesh Jain (Jira)
Lokesh Jain created HDFS-15392:
--

 Summary: DistrbutedFileSystem#concat api can create large number 
of small blocks
 Key: HDFS-15392
 URL: https://issues.apache.org/jira/browse/HDFS-15392
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Lokesh Jain


DistrbutedFileSystem#concat moves blocks from source to target. If the api is 
repeatedly used on small files it can create large number of small blocks in 
the target file. The Jira aims to optimize the api to avoid the issue of small 
blocks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-15201) SnapshotCounter hits MaxSnapshotID limit

2020-03-24 Thread Lokesh Jain (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDFS-15201.

Resolution: Fixed

> SnapshotCounter hits MaxSnapshotID limit
> 
>
> Key: HDFS-15201
> URL: https://issues.apache.org/jira/browse/HDFS-15201
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: snapshots
>Reporter: Karthik Palanisamy
>Assignee: Karthik Palanisamy
>Priority: Major
>
> Users reported that they are unable to take HDFS snapshots and their 
> snapshotCounter hits MaxSnapshotID limit. MaxSnapshotID limit is 16777215.
> {code:java}
> SnapshotManager.java
> private static final int SNAPSHOT_ID_BIT_WIDTH = 24;
> /**
>  * Returns the maximum allowable snapshot ID based on the bit width of the
>  * snapshot ID.
>  *
>  * @return maximum allowable snapshot ID.
>  */
>  public int getMaxSnapshotID() {
>  return ((1 << SNAPSHOT_ID_BIT_WIDTH) - 1);
> }
> {code}
>  
> I think, SNAPSHOT_ID_BIT_WIDTH is too low. May be good idea to increase 
> SNAPSHOT_ID_BIT_WIDTH to 31? to aline with our CURRENT_STATE_ID limit 
> (Integer.MAX_VALUE - 1).
>  
> {code:java}
> /**
>  * This id is used to indicate the current state (vs. snapshots)
>  */
> public static final int CURRENT_STATE_ID = Integer.MAX_VALUE - 1;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-2347) XCeiverClientGrpc's parallel use leads to NPE

2019-10-30 Thread Lokesh Jain (Jira)


 [ 
https://issues.apache.org/jira/browse/HDDS-2347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-2347.
---
Fix Version/s: 0.5.0
   Resolution: Fixed

> XCeiverClientGrpc's parallel use leads to NPE
> -
>
> Key: HDDS-2347
> URL: https://issues.apache.org/jira/browse/HDDS-2347
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Client
>Reporter: Istvan Fajth
>Assignee: Istvan Fajth
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 0.5.0
>
> Attachments: changes.diff, logs.txt
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> This issue came up when testing Hive with ORC tables on Ozone storage 
> backend, I so far I could not reproduce it locally within a JUnit test but 
> the issue.
> I am attaching a diff file that shows what logging I have added in 
> XCevierClientGrpc and in KeyInputStream to get the results that made me 
> arrive to the following understanding of the scenario:
> - Hive starts a couple of threads to work on the table data during query 
> execution
> - There is one RPCClient that is being used by these threads
> - The threads are opening different stream to read from the same key in ozone
> - The InputStreams internally are using the same XCeiverClientGrpc
> - XCeiverClientGrpc throws the following NPE intermittently:
> {code}
> Caused by: java.lang.NullPointerException
> at 
> org.apache.hadoop.hdds.scm.XceiverClientGrpc.sendCommandAsync(XceiverClientGrpc.java:398)
> at 
> org.apache.hadoop.hdds.scm.XceiverClientGrpc.sendCommandWithRetry(XceiverClientGrpc.java:295)
> at 
> org.apache.hadoop.hdds.scm.XceiverClientGrpc.sendCommandWithTraceIDAndRetry(XceiverClientGrpc.java:259)
> at 
> org.apache.hadoop.hdds.scm.XceiverClientGrpc.sendCommand(XceiverClientGrpc.java:242)
> at 
> org.apache.hadoop.hdds.scm.storage.ContainerProtocolCalls.getBlock(ContainerProtocolCalls.java:118)
> at 
> org.apache.hadoop.hdds.scm.storage.BlockInputStream.getChunkInfos(BlockInputStream.java:169)
> at 
> org.apache.hadoop.hdds.scm.storage.BlockInputStream.initialize(BlockInputStream.java:118)
> at 
> org.apache.hadoop.hdds.scm.storage.BlockInputStream.read(BlockInputStream.java:224)
> at 
> org.apache.hadoop.ozone.client.io.KeyInputStream.read(KeyInputStream.java:173)
> at 
> org.apache.hadoop.fs.ozone.OzoneFSInputStream.read(OzoneFSInputStream.java:52)
> at org.apache.hadoop.fs.FSInputStream.read(FSInputStream.java:75)
> at 
> org.apache.hadoop.fs.FSInputStream.readFully(FSInputStream.java:121)
> at 
> org.apache.hadoop.fs.FSDataInputStream.readFully(FSDataInputStream.java:112)
> at org.apache.orc.impl.ReaderImpl.extractFileTail(ReaderImpl.java:555)
> at org.apache.orc.impl.ReaderImpl.(ReaderImpl.java:370)
> at 
> org.apache.hadoop.hive.ql.io.orc.ReaderImpl.(ReaderImpl.java:61)
> at 
> org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader(OrcFile.java:105)
> at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:1708)
> at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal(OrcInputFormat.java:1596)
> at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2900(OrcInputFormat.java:1383)
> at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1568)
> at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run(OrcInputFormat.java:1565)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1876)
> at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1565)
> at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:1383)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> {code}
> I have two proposals to fix this issue, one is the easy answer to put 
> synchronization to the XCeiverClientGrpc code, the other one is a bit more 
> complicated, let me explain below.
> Naively I would assume that when I get a client SPI instance from 
> XCeiverClientManager, that instance is ready to use. In fact it is not, and 
> when the user of the SPI instance sends the first request that is the point 
> when the client gets essentially ready. Now if we put synchronization to this 
> code, that is the easy solution, but my pragmatic half screams for a better 
> solution, that ensures that 

[jira] [Created] (HDDS-2342) ContainerStateMachine$chunkExecutor threads hold onto native memory

2019-10-21 Thread Lokesh Jain (Jira)
Lokesh Jain created HDDS-2342:
-

 Summary: ContainerStateMachine$chunkExecutor threads hold onto 
native memory
 Key: HDDS-2342
 URL: https://issues.apache.org/jira/browse/HDDS-2342
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain


In a heap dump many threads in ContainerStateMachine$chunkExecutor holds onto 
native memory in the ThreadLocal map. Every such thread holds onto chunk worth 
of DirectByteBuffer. Since these threads are involved in write and read chunk 
operations, the JVM allocates chunk (16MB) worth of DirectByteBuffer in the 
ThreadLocalMap for every thread involved in IO. Also the native memory would 
not be GC'ed as long as the thread is alive.

It would be better to reduce the default number of chunk executor threads and 
have them in proportion to number of disks on the datanode.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-2332) BlockOutputStream#waitOnFlushFutures blocks on putBlock combined future

2019-10-18 Thread Lokesh Jain (Jira)
Lokesh Jain created HDDS-2332:
-

 Summary: BlockOutputStream#waitOnFlushFutures blocks on putBlock 
combined future
 Key: HDDS-2332
 URL: https://issues.apache.org/jira/browse/HDDS-2332
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Client
Reporter: Lokesh Jain


BlockOutputStream blocks on waitOnFlushFutures call. Two jstacks show that the 
thread is blocked on the same condition.
{code:java}
2019-10-18 06:30:38
Full thread dump Java HotSpot(TM) 64-Bit Server VM (25.141-b15 mixed mode):
"main" #1 prio=5 os_prio=0 tid=0x7fbea001a800 nid=0x2a56 waiting on 
condition [0x7fbea96d6000]
   java.lang.Thread.State: WAITING (parking)
at sun.misc.Unsafe.park(Native Method)
- parking to wait for  <0xe4739888> (a 
java.util.concurrent.CompletableFuture$Signaller)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at 
java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1693)
at 
java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323)
at 
java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1729)
at 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.waitOnFlushFutures(BlockOutputStream.java:518)
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.handleFlush(BlockOutputStream.java:481)
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.close(BlockOutputStream.java:496)
at 
org.apache.hadoop.ozone.client.io.BlockOutputStreamEntry.close(BlockOutputStreamEntry.java:143)
at 
org.apache.hadoop.ozone.client.io.KeyOutputStream.handleFlushOrClose(KeyOutputStream.java:439)
at 
org.apache.hadoop.ozone.client.io.KeyOutputStream.handleWrite(KeyOutputStream.java:232)
at 
org.apache.hadoop.ozone.client.io.KeyOutputStream.write(KeyOutputStream.java:190)
at 
org.apache.hadoop.fs.ozone.OzoneFSOutputStream.write(OzoneFSOutputStream.java:46)
at 
org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:57)
at java.io.DataOutputStream.write(DataOutputStream.java:107)
- locked <0xa6a75930> (a 
org.apache.hadoop.fs.FSDataOutputStream)
at 
org.apache.hadoop.examples.terasort.TeraOutputFormat$TeraRecordWriter.write(TeraOutputFormat.java:77)
- locked <0xa6a75918> (a 
org.apache.hadoop.examples.terasort.TeraOutputFormat$TeraRecordWriter)
at 
org.apache.hadoop.examples.terasort.TeraOutputFormat$TeraRecordWriter.write(TeraOutputFormat.java:64)
at 
org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.write(MapTask.java:670)
at 
org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(TaskInputOutputContextImpl.java:89)
at 
org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.write(WrappedMapper.java:112)
at 
org.apache.hadoop.examples.terasort.TeraGen$SortGenMapper.map(TeraGen.java:230)
at 
org.apache.hadoop.examples.terasort.TeraGen$SortGenMapper.map(TeraGen.java:203)
at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146)
at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:799)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:347)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:174)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1876)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:168)


2019-10-18 07:02:50
Full thread dump Java HotSpot(TM) 64-Bit Server VM (25.141-b15 mixed mode):
"main" #1 prio=5 os_prio=0 tid=0x7fbea001a800 nid=0x2a56 waiting on 
condition [0x7fbea96d6000]
   java.lang.Thread.State: WAITING (parking)
at sun.misc.Unsafe.park(Native Method)
- parking to wait for  <0xe4739888> (a 
java.util.concurrent.CompletableFuture$Signaller)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at 
java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1693)
at 
java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323)
at 
java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1729)
at 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.waitOnFlushFutures(BlockOutputStream.java:518)
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.handleFlush(BlockOutputStream.java:481)
at 

[jira] [Created] (HDDS-2299) BlockManager should allocate a block in excluded pipelines if none other left

2019-10-14 Thread Lokesh Jain (Jira)
Lokesh Jain created HDDS-2299:
-

 Summary: BlockManager should allocate a block in excluded 
pipelines if none other left
 Key: HDDS-2299
 URL: https://issues.apache.org/jira/browse/HDDS-2299
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain


In SCM, BlockManager#allocateBlock does not allocate a block in the excluded 
pipelines or datanodes if requested by the client. But there can be cases where 
excluded pipelines are the only pipelines left. In such a case SCM should 
allocate a block in such pipelines and return to the client. The client can 
choose to use or discard the block.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-2189) Datanode should send PipelineAction on RaftServer failure

2019-09-26 Thread Lokesh Jain (Jira)
Lokesh Jain created HDDS-2189:
-

 Summary: Datanode should send PipelineAction on RaftServer failure
 Key: HDDS-2189
 URL: https://issues.apache.org/jira/browse/HDDS-2189
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain


{code:java}
2019-09-26 08:03:07,152 ERROR 
org.apache.ratis.server.raftlog.segmented.SegmentedRaftLogWorker: 
664c4e90-08f3-46c9-a073-c93ef2a55da3@group-93F633896F08-SegmentedRaftLogWorker 
hit exception
java.lang.OutOfMemoryError: Direct buffer memory
at java.nio.Bits.reserveMemory(Bits.java:694)
at java.nio.DirectByteBuffer.(DirectByteBuffer.java:123)
at java.nio.ByteBuffer.allocateDirect(ByteBuffer.java:311)
at 
org.apache.ratis.server.raftlog.segmented.BufferedWriteChannel.(BufferedWriteChannel.java:41)
at 
org.apache.ratis.server.raftlog.segmented.SegmentedRaftLogOutputStream.(SegmentedRaftLogOutputStream.java:72)
at 
org.apache.ratis.server.raftlog.segmented.SegmentedRaftLogWorker$StartLogSegment.execute(SegmentedRaftLogWorker.java:566)
at 
org.apache.ratis.server.raftlog.segmented.SegmentedRaftLogWorker.run(SegmentedRaftLogWorker.java:289)
at java.lang.Thread.run(Thread.java:748)
2019-09-26 08:03:07,155 INFO org.apache.ratis.server.impl.RaftServerImpl: 
664c4e90-08f3-46c9-a073-c93ef2a55da3@group-93F633896F08: shutdown
{code}
On RaftServer shutdown datanode should send a PipelineAction denoting that the 
pipeline has been closed exceptionally in the datanode.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-2048) State check during container state transition in datanode should be lock protected

2019-08-28 Thread Lokesh Jain (Jira)
Lokesh Jain created HDDS-2048:
-

 Summary: State check during container state transition in datanode 
should be lock protected
 Key: HDDS-2048
 URL: https://issues.apache.org/jira/browse/HDDS-2048
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently container state checks during state transition are not lock protected 
in KeyValueHandler. These can cause invalid state transitions.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1959) Decrement purge interval for Ratis logs

2019-08-13 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1959:
-

 Summary: Decrement purge interval for Ratis logs
 Key: HDDS-1959
 URL: https://issues.apache.org/jira/browse/HDDS-1959
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain


Currently purge interval for ratis log("ozone.om.ratis.log.purge.gap") is set 
at 100. The Jira aims to reduce the interval and set it to 10.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14692) Upload button should not encode complete url

2019-08-01 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDFS-14692:
--

 Summary: Upload button should not encode complete url
 Key: HDFS-14692
 URL: https://issues.apache.org/jira/browse/HDFS-14692
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


explorer.js#modal-upload-file-button currently does not work with knox. The 
function encodes the complete url and thus creates a malformed url. This leads 
to an error while uploading the file.

Example of malformed url - 
"https%3A//127.0.0.1%3A/gateway/default/webhdfs/v1/app-logs/BUILDING.txt?op=CREATE=true"



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1824) IllegalArgumentException in NetworkTopologyImpl causes SCM to shutdown

2019-07-18 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1824:
-

 Summary: IllegalArgumentException in NetworkTopologyImpl causes 
SCM to shutdown
 Key: HDDS-1824
 URL: https://issues.apache.org/jira/browse/HDDS-1824
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain


 

 
{code:java}
2019-07-18 02:22:18,005 ERROR 
org.apache.hadoop.hdds.scm.container.ReplicationManager: Exception in 
Replication Monitor Thread.
java.lang.IllegalArgumentException: Affinity node /default-rack/10.17.213.25 is 
not a member of topology
at 
org.apache.hadoop.hdds.scm.net.NetworkTopologyImpl.checkAffinityNode(NetworkTopologyImpl.java:780)
at 
org.apache.hadoop.hdds.scm.net.NetworkTopologyImpl.chooseRandom(NetworkTopologyImpl.java:408)
at 
org.apache.hadoop.hdds.scm.container.placement.algorithms.SCMContainerPlacementRackAware.chooseNode(SCMContainerPlacementRackAware.java:242)
at 
org.apache.hadoop.hdds.scm.container.placement.algorithms.SCMContainerPlacementRackAware.chooseDatanodes(SCMContainerPlacementRackAware.java:168)
at 
org.apache.hadoop.hdds.scm.container.ReplicationManager.handleUnderReplicatedContainer(ReplicationManager.java:487)
at 
org.apache.hadoop.hdds.scm.container.ReplicationManager.processContainer(ReplicationManager.java:293)
at 
java.util.concurrent.ConcurrentHashMap$KeySetView.forEach(ConcurrentHashMap.java:4649)
at java.util.Collections$UnmodifiableCollection.forEach(Collections.java:1080)
at 
org.apache.hadoop.hdds.scm.container.ReplicationManager.run(ReplicationManager.java:205)
at java.lang.Thread.run(Thread.java:745)
2019-07-18 02:22:18,008 INFO org.apache.hadoop.util.ExitUtil: Exiting with 
status 1: java.lang.IllegalArgumentException: Affinity node 
/default-rack/10.17.213.25 is not a member of topology
2019-07-18 02:22:18,010 INFO 
org.apache.hadoop.hdds.scm.server.StorageContainerManagerStarter: SHUTDOWN_MSG:
{code}
 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1816) ContainerStateMachine should limit number of pending apply transactions

2019-07-17 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1816:
-

 Summary: ContainerStateMachine should limit number of pending 
apply transactions
 Key: HDDS-1816
 URL: https://issues.apache.org/jira/browse/HDDS-1816
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


ContainerStateMachine should limit number of pending apply transactions in 
order to avoid excessive heap usage by the pending transactions.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1767) ContainerStateMachine should have its own executors for executing applyTransaction calls

2019-07-05 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1767:
-

 Summary: ContainerStateMachine should have its own executors for 
executing applyTransaction calls
 Key: HDDS-1767
 URL: https://issues.apache.org/jira/browse/HDDS-1767
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently ContainerStateMachine uses the executors provided by 
XceiverServerRatis for executing applyTransaction calls. This would result in 
two or more ContainerStateMachine to share the same set of executors. Delay or 
load in one ContainerStateMachine would adversely affect the performance of 
other state machines in such a case. It is better to have separate set of 
executors for each ContainerStateMachine.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1766) ContainerStateMachine is unable to increment lastAppliedIndex

2019-07-05 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1766:
-

 Summary: ContainerStateMachine is unable to increment 
lastAppliedIndex
 Key: HDDS-1766
 URL: https://issues.apache.org/jira/browse/HDDS-1766
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain


ContainerStateMachine#updateLastApplied currently updates the 
lastAppliedTermIndex using applyTransactionCompletionMap. There are null 
entries in the applyTransactionCompletionMap causing the lastAppliedIndex to 
not be incremented.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1750) Add block allocation metric for pipelines in SCM

2019-07-02 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1750:
-

 Summary: Add block allocation metric for pipelines in SCM
 Key: HDDS-1750
 URL: https://issues.apache.org/jira/browse/HDDS-1750
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


This Jira aims to add block allocation metrics for pipelines in SCM. This would 
help in determining the distribution of block allocations among various 
pipelines in SCM.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1626) Optimize allocateBlock for cases when excludeList is provided

2019-06-02 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1626:
-

 Summary: Optimize allocateBlock for cases when excludeList is 
provided
 Key: HDDS-1626
 URL: https://issues.apache.org/jira/browse/HDDS-1626
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


This Jira aims to optimize allocateBlock for cases when excludeList is 
provided. This includes the case when excludeList is empty or the cases when it 
is not empty.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-12735) Make ContainerStateMachine#applyTransaction async

2019-05-11 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDFS-12735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDFS-12735.

Resolution: Duplicate

> Make ContainerStateMachine#applyTransaction async
> -
>
> Key: HDFS-12735
> URL: https://issues.apache.org/jira/browse/HDFS-12735
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Lokesh Jain
>Assignee: Lokesh Jain
>Priority: Major
>  Labels: performance
> Attachments: HDFS-12735-HDFS-7240.000.patch, 
> HDFS-12735-HDFS-7240.001.patch, HDFS-12735-HDFS-7240.002.patch
>
>
> Currently ContainerStateMachine#applyTransaction makes a synchronous call to 
> dispatch client requests. Idea is to have a thread pool which dispatches 
> client requests and returns a CompletableFuture.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1503) Reduce garbage generated by non-netty threads in datanode ratis server

2019-05-08 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1503:
-

 Summary: Reduce garbage generated by non-netty threads in datanode 
ratis server
 Key: HDDS-1503
 URL: https://issues.apache.org/jira/browse/HDDS-1503
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain


We use GRPC protocol for rpc communication in Ratis. By default thread caches 
are generated even for non-netty threads. This Jira aims to add a default JVM 
parameter for disabling thread caches for non-netty threads in datanode ratis 
server.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1481) Cleanup BasicOzoneFileSystem#mkdir

2019-05-01 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1481:
-

 Summary: Cleanup BasicOzoneFileSystem#mkdir
 Key: HDDS-1481
 URL: https://issues.apache.org/jira/browse/HDDS-1481
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Filesystem
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently BasicOzoneFileSystem#mkdir does not have the optimizations made in 
HDDS-1300. The changes for this function were missed in HDDS-1460.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1461) Optimize listStatus api in OzoneFileStatus

2019-04-24 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1461:
-

 Summary: Optimize listStatus api in OzoneFileStatus
 Key: HDDS-1461
 URL: https://issues.apache.org/jira/browse/HDDS-1461
 Project: Hadoop Distributed Data Store
  Issue Type: Sub-task
  Components: Ozone Filesystem, Ozone Manager
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently in listStatus we make multiple getFileStatus calls. This can be 
optimized by converting to a single rpc call for listStatus.

Also currently listStatus has to traverse a directory recursively in order to 
list its immediate children. This happens because in OzoneManager all the 
metadata is stored in rocksdb sorted on keynames. The Jira also aims to fix 
this by using seek api provided by rocksdb.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1460) Add the optmizations of HDDS-1300 to BasicOzoneFileSystem

2019-04-24 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1460:
-

 Summary: Add the optmizations of HDDS-1300 to BasicOzoneFileSystem
 Key: HDDS-1460
 URL: https://issues.apache.org/jira/browse/HDDS-1460
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Some of the optimizations made in HDDS-1300 were reverted in HDDS-1333. This 
Jira aims to bring back those optimizations.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-1405) ITestOzoneContractCreate is failing

2019-04-09 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1405?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-1405.
---
Resolution: Resolved

> ITestOzoneContractCreate is failing
> ---
>
> Key: HDDS-1405
> URL: https://issues.apache.org/jira/browse/HDDS-1405
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>Reporter: Lokesh Jain
>Assignee: Lokesh Jain
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> ITestOzoneContractCreate and ITestOzoneContractMkdir are failing with 
> FileAlreadyExistsException. The issue is around the file imported in 
> BasicOzoneClientAdapterImpl. The class needs to import 
> org.apache.hadoop.fs.FileAlreadyExistsException but currently imports 
> java.nio.file.FileAlreadyExistsException.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1405) ITestOzoneContractCreate is failing

2019-04-09 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1405:
-

 Summary: ITestOzoneContractCreate is failing
 Key: HDDS-1405
 URL: https://issues.apache.org/jira/browse/HDDS-1405
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


ITestOzoneContractCreate and ITestOzoneContractMkdir are failing with 
FileAlreadyExistsException. The issue is around the file imported in 
BasicOzoneClientAdapterImpl. The class needs to import 
org.apache.hadoop.fs.FileAlreadyExistsException but currently imports 
java.nio.file.FileAlreadyExistsException.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Reopened] (HDDS-1134) OzoneFileSystem#create should allocate alteast one block for future writes.

2019-03-29 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain reopened HDDS-1134:
---

Reopening issue as it was not fixed in HDDS-1300.

> OzoneFileSystem#create should allocate alteast one block for future writes.
> ---
>
> Key: HDDS-1134
> URL: https://issues.apache.org/jira/browse/HDDS-1134
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: Ozone Manager
>Affects Versions: 0.4.0
>Reporter: Mukul Kumar Singh
>Assignee: Mukul Kumar Singh
>Priority: Major
> Attachments: HDDS-1134.001.patch
>
>
> While opening a new key, OM should at least allocate one block for the key, 
> this should be done in case the client is not sure about the number of block. 
> However for users of OzoneFS, if the key is being created for a directory, 
> then no blocks should be allocated.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1343) TestNodeFailure times out intermittently

2019-03-27 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1343:
-

 Summary: TestNodeFailure times out intermittently
 Key: HDDS-1343
 URL: https://issues.apache.org/jira/browse/HDDS-1343
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain


TestNodeFailure times out while waiting for cluster to be ready. This is done 
in cluster setup.
{code:java}
java.lang.Thread.State: WAITING (on object monitor)
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at 
java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:403)
at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)


at 
org.apache.hadoop.test.GenericTestUtils.waitFor(GenericTestUtils.java:389)
at 
org.apache.hadoop.ozone.MiniOzoneClusterImpl.waitForClusterToBeReady(MiniOzoneClusterImpl.java:140)
at 
org.apache.hadoop.hdds.scm.pipeline.TestNodeFailure.init(TestNodeFailure.java:74)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
at 
org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
{code}
5 datanodes out of 6 are able to heartbeat in the test result 
[https://builds.apache.org/job/PreCommit-HDDS-Build/2582/testReport/].



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1342) TestOzoneManagerHA#testOMProxyProviderFailoverOnConnectionFailure fails intermittently

2019-03-27 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1342:
-

 Summary: 
TestOzoneManagerHA#testOMProxyProviderFailoverOnConnectionFailure fails 
intermittently
 Key: HDDS-1342
 URL: https://issues.apache.org/jira/browse/HDDS-1342
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain


The test fails intermittently. The link to the test report can be found below.

[https://builds.apache.org/job/PreCommit-HDDS-Build/2582/testReport/]
{code:java}
java.net.ConnectException: Call From ea902c1cb730/172.17.0.3 to localhost:10174 
failed on connection exception: java.net.ConnectException: Connection refused; 
For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:831)
at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:755)
at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1515)
at org.apache.hadoop.ipc.Client.call(Client.java:1457)
at org.apache.hadoop.ipc.Client.call(Client.java:1367)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
at com.sun.proxy.$Proxy34.submitRequest(Unknown Source)
at sun.reflect.GeneratedMethodAccessor30.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
at com.sun.proxy.$Proxy34.submitRequest(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.hadoop.hdds.tracing.TraceAllMethod.invoke(TraceAllMethod.java:66)
at com.sun.proxy.$Proxy34.submitRequest(Unknown Source)
at 
org.apache.hadoop.ozone.om.protocolPB.OzoneManagerProtocolClientSideTranslatorPB.submitRequest(OzoneManagerProtocolClientSideTranslatorPB.java:310)
at 
org.apache.hadoop.ozone.om.protocolPB.OzoneManagerProtocolClientSideTranslatorPB.createVolume(OzoneManagerProtocolClientSideTranslatorPB.java:343)
at 
org.apache.hadoop.ozone.client.rpc.RpcClient.createVolume(RpcClient.java:275)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.hadoop.ozone.client.OzoneClientInvocationHandler.invoke(OzoneClientInvocationHandler.java:54)
at com.sun.proxy.$Proxy86.createVolume(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.hadoop.hdds.tracing.TraceAllMethod.invoke(TraceAllMethod.java:66)
at com.sun.proxy.$Proxy86.createVolume(Unknown Source)
at 
org.apache.hadoop.ozone.client.ObjectStore.createVolume(ObjectStore.java:100)
at 
org.apache.hadoop.ozone.om.TestOzoneManagerHA.createVolumeTest(TestOzoneManagerHA.java:162)
at 
org.apache.hadoop.ozone.om.TestOzoneManagerHA.testOMProxyProviderFailoverOnConnectionFailure(TestOzoneManagerHA.java:237)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 

[jira] [Created] (HDDS-1341) TestContainerReplication#testContainerReplication fails intermittently

2019-03-27 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1341:
-

 Summary: TestContainerReplication#testContainerReplication fails 
intermittently
 Key: HDDS-1341
 URL: https://issues.apache.org/jira/browse/HDDS-1341
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain


The test fails intermittently. The link to the test report can be found below.

https://builds.apache.org/job/PreCommit-HDDS-Build/2582/testReport/
{code:java}
java.lang.AssertionError: Container is not replicated to the destination 
datanode
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertNotNull(Assert.java:621)
at 
org.apache.hadoop.ozone.container.TestContainerReplication.testContainerReplication(TestContainerReplication.java:139)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74)
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1300) Optimize non-recursive ozone filesystem apis

2019-03-18 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1300:
-

 Summary: Optimize non-recursive ozone filesystem apis
 Key: HDDS-1300
 URL: https://issues.apache.org/jira/browse/HDDS-1300
 Project: Hadoop Distributed Data Store
  Issue Type: Sub-task
  Components: Ozone Filesystem, Ozone Manager
Reporter: Lokesh Jain
Assignee: Lokesh Jain


This Jira aims to optimise non recursive apis in ozone file system. The Jira 
would add support for such apis in Ozone manager in order to reduce the number 
of rpc calls to Ozone Manager.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1301) Optimize recursive ozone filesystem apis

2019-03-18 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1301:
-

 Summary: Optimize recursive ozone filesystem apis
 Key: HDDS-1301
 URL: https://issues.apache.org/jira/browse/HDDS-1301
 Project: Hadoop Distributed Data Store
  Issue Type: Sub-task
Reporter: Lokesh Jain
Assignee: Lokesh Jain


This Jira aims to optimise recursive apis in ozone file system. These are the 
apis which have a recursive flag which requires an operation to be performed on 
all the children of the directory. The Jira would add support for recursive 
apis in Ozone manager in order to reduce the number of rpc calls to Ozone 
Manager.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Reopened] (HDDS-1248) TestSecureOzoneRpcClient fails intermittently

2019-03-11 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain reopened HDDS-1248:
---

The test calls BlockTokenIdentifier#setTestStub(true) in 

TestSecureOzoneRpcClient#testKeyOpFailureWithoutBlockToken. Since testStub is 
true all the concurrently running tests fail with Block token verification 
failed exception.

> TestSecureOzoneRpcClient fails intermittently
> -
>
> Key: HDDS-1248
> URL: https://issues.apache.org/jira/browse/HDDS-1248
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>Reporter: Lokesh Jain
>Priority: Major
> Fix For: 0.4.0
>
>
>  
> TestSecureOzoneRpcClient fails intermittently with the following exception.
> {code:java}
> java.io.IOException: Unexpected Storage Container Exception: 
> java.util.concurrent.ExecutionException: 
> java.util.concurrent.CompletionException: 
> org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException:
>  Block token verification failed. Fail to find any token (empty or null.
>   at 
> org.apache.hadoop.hdds.scm.storage.BlockOutputStream.handleFullBuffer(BlockOutputStream.java:338)
>   at 
> org.apache.hadoop.hdds.scm.storage.BlockOutputStream.write(BlockOutputStream.java:238)
>   at 
> org.apache.hadoop.ozone.client.io.BlockOutputStreamEntry.write(BlockOutputStreamEntry.java:131)
>   at 
> org.apache.hadoop.ozone.client.io.KeyOutputStream.handleWrite(KeyOutputStream.java:310)
>   at 
> org.apache.hadoop.ozone.client.io.KeyOutputStream.write(KeyOutputStream.java:271)
>   at 
> org.apache.hadoop.ozone.client.io.OzoneOutputStream.write(OzoneOutputStream.java:49)
>   at 
> org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.uploadPart(TestOzoneRpcClientAbstract.java:2188)
>   at 
> org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.doMultipartUpload(TestOzoneRpcClientAbstract.java:2131)
>   at 
> org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.testMultipartUpload(TestOzoneRpcClientAbstract.java:1721)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
> Caused by: java.util.concurrent.ExecutionException: 
> java.util.concurrent.CompletionException: 
> org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException:
>  Block token verification failed. Fail to find any token (empty or null.
>   at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
>   at 
> 

[jira] [Resolved] (HDDS-1248) TestSecureOzoneRpcClient fails intermittently

2019-03-11 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-1248.
---
Resolution: Duplicate

> TestSecureOzoneRpcClient fails intermittently
> -
>
> Key: HDDS-1248
> URL: https://issues.apache.org/jira/browse/HDDS-1248
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>Reporter: Lokesh Jain
>Priority: Major
> Fix For: 0.4.0
>
>
>  
> TestSecureOzoneRpcClient fails intermittently with the following exception.
> {code:java}
> java.io.IOException: Unexpected Storage Container Exception: 
> java.util.concurrent.ExecutionException: 
> java.util.concurrent.CompletionException: 
> org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException:
>  Block token verification failed. Fail to find any token (empty or null.
>   at 
> org.apache.hadoop.hdds.scm.storage.BlockOutputStream.handleFullBuffer(BlockOutputStream.java:338)
>   at 
> org.apache.hadoop.hdds.scm.storage.BlockOutputStream.write(BlockOutputStream.java:238)
>   at 
> org.apache.hadoop.ozone.client.io.BlockOutputStreamEntry.write(BlockOutputStreamEntry.java:131)
>   at 
> org.apache.hadoop.ozone.client.io.KeyOutputStream.handleWrite(KeyOutputStream.java:310)
>   at 
> org.apache.hadoop.ozone.client.io.KeyOutputStream.write(KeyOutputStream.java:271)
>   at 
> org.apache.hadoop.ozone.client.io.OzoneOutputStream.write(OzoneOutputStream.java:49)
>   at 
> org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.uploadPart(TestOzoneRpcClientAbstract.java:2188)
>   at 
> org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.doMultipartUpload(TestOzoneRpcClientAbstract.java:2131)
>   at 
> org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.testMultipartUpload(TestOzoneRpcClientAbstract.java:1721)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
> Caused by: java.util.concurrent.ExecutionException: 
> java.util.concurrent.CompletionException: 
> org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException:
>  Block token verification failed. Fail to find any token (empty or null.
>   at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
>   at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
>   at 
> org.apache.hadoop.hdds.scm.storage.BlockOutputStream.waitOnFlushFutures(BlockOutputStream.java:543)
>   at 
> 

[jira] [Created] (HDDS-1248) TestSecureOzoneRpcClient fails intermittently

2019-03-11 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1248:
-

 Summary: TestSecureOzoneRpcClient fails intermittently
 Key: HDDS-1248
 URL: https://issues.apache.org/jira/browse/HDDS-1248
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
 Fix For: 0.4.0


 

TestSecureOzoneRpcClient fails intermittently with the following exception.
{code:java}
java.io.IOException: Unexpected Storage Container Exception: 
java.util.concurrent.ExecutionException: 
java.util.concurrent.CompletionException: 
org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException: 
Block token verification failed. Fail to find any token (empty or null.
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.handleFullBuffer(BlockOutputStream.java:338)
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.write(BlockOutputStream.java:238)
at 
org.apache.hadoop.ozone.client.io.BlockOutputStreamEntry.write(BlockOutputStreamEntry.java:131)
at 
org.apache.hadoop.ozone.client.io.KeyOutputStream.handleWrite(KeyOutputStream.java:310)
at 
org.apache.hadoop.ozone.client.io.KeyOutputStream.write(KeyOutputStream.java:271)
at 
org.apache.hadoop.ozone.client.io.OzoneOutputStream.write(OzoneOutputStream.java:49)
at 
org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.uploadPart(TestOzoneRpcClientAbstract.java:2188)
at 
org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.doMultipartUpload(TestOzoneRpcClientAbstract.java:2131)
at 
org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.testMultipartUpload(TestOzoneRpcClientAbstract.java:1721)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
at 
org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
Caused by: java.util.concurrent.ExecutionException: 
java.util.concurrent.CompletionException: 
org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException: 
Block token verification failed. Fail to find any token (empty or null.
at 
java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
at 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.waitOnFlushFutures(BlockOutputStream.java:543)
at 
org.apache.hadoop.hdds.scm.storage.BlockOutputStream.handleFullBuffer(BlockOutputStream.java:333)
... 35 more
Caused by: java.util.concurrent.CompletionException: 
org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException: 
Block token verification failed. Fail to find any token (empty or null.
at 

[jira] [Created] (HDDS-1221) Introduce fine grained lock in Ozone Manager for key operations

2019-03-05 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1221:
-

 Summary: Introduce fine grained lock in Ozone Manager for key 
operations
 Key: HDDS-1221
 URL: https://issues.apache.org/jira/browse/HDDS-1221
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently ozone manager acquires bucket lock for key operations in OM. We can 
introduce fine grained lock for key operations in ozone manager. This would 
help in increasing throughput for key operations in a bucket.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1220) KeyManager#openKey should release the bucket lock before doing an allocateBlock

2019-03-05 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1220:
-

 Summary: KeyManager#openKey should release the bucket lock before 
doing an allocateBlock
 Key: HDDS-1220
 URL: https://issues.apache.org/jira/browse/HDDS-1220
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently KeyManager#openKey makes an allocateBlock call without releasing the 
bucket lock. Since allocateBlock requires a rpc connection to SCM, the handler 
thread in OM would hold the bucket lock until rpc is complete. Since 
allocateBlock call does not require a bucket lock to be held, allocateBlock 
call can be made after releasing the bucket lock.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1208) ContainerStateMachine should set chunk data as state machine data for ratis

2019-03-01 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1208:
-

 Summary: ContainerStateMachine should set chunk data as state 
machine data for ratis
 Key: HDDS-1208
 URL: https://issues.apache.org/jira/browse/HDDS-1208
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently ContainerStateMachine sets ContainerCommandRequestProto as state 
machine data. This requires converting the ContainerCommandRequestProto to a 
bytestring which leads to redundant buffer copy in case of write chunk request. 
This can be avoided by setting the chunk data as the state machine data for a 
log entry in ratis.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-1140) TestSCMChillModeManager is failing with NullPointerException

2019-02-27 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-1140.
---
Resolution: Duplicate

> TestSCMChillModeManager is failing with NullPointerException
> 
>
> Key: HDDS-1140
> URL: https://issues.apache.org/jira/browse/HDDS-1140
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>  Components: SCM
>Affects Versions: 0.4.0
>Reporter: Mukul Kumar Singh
>Assignee: Lokesh Jain
>Priority: Major
>
> TestSCMChillModeManager is failing with the following exception
> {code}
> [ERROR] 
> testDisableChillMode(org.apache.hadoop.hdds.scm.chillmode.TestSCMChillModeManager)
>   Time elapsed: 0.012 s  <<< ERROR!
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.utils.Scheduler.scheduleWithFixedDelay(Scheduler.java:78)
>   at 
> org.apache.hadoop.hdds.scm.pipeline.RatisPipelineUtils.scheduleFixedIntervalPipelineCreator(RatisPipelineUtils.java:211)
>   at 
> org.apache.hadoop.hdds.scm.chillmode.SCMChillModeManager.exitChillMode(SCMChillModeManager.java:137)
>   at 
> org.apache.hadoop.hdds.scm.chillmode.SCMChillModeManager.(SCMChillModeManager.java:93)
>   at 
> org.apache.hadoop.hdds.scm.chillmode.TestSCMChillModeManager.testDisableChillMode(TestSCMChillModeManager.java:134)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1171) Add benchmark for OM and OM client in Genesis

2019-02-25 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1171:
-

 Summary: Add benchmark for OM and OM client in Genesis
 Key: HDDS-1171
 URL: https://issues.apache.org/jira/browse/HDDS-1171
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


This Jira aims to add benchmark for OM and OM client in Genesis.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1130) Make BenchMarkBlockManager multi-threaded

2019-02-18 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1130:
-

 Summary: Make BenchMarkBlockManager multi-threaded
 Key: HDDS-1130
 URL: https://issues.apache.org/jira/browse/HDDS-1130
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


Currently BenchMarkBlockManager is run by a single thread. We can make it 
multi-threaded in order to have a better understanding of allocateBlock call 
performance.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1106) Introduce queryMap in PipelineManager

2019-02-14 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1106:
-

 Summary: Introduce queryMap in PipelineManager
 Key: HDDS-1106
 URL: https://issues.apache.org/jira/browse/HDDS-1106
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


In Genesis benchmark for block allocation it was found that 
BlockManager#allocateBlock call was very slow for higher number of pipelines. 
This happens because allocateBlock call gets list of pipelines with a 
particular replication type, replication factor and state. This list is 
calculated by traversing the entries of a map. This Jira aims to optimize the 
call by introducing query map in Pipeline Manager.

The pipeline manager would keep a maintain a list of pipelines for every query 
type i.e. for every replication type, replication factor.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1097) Add genesis benchmark for BlockManager#allocateBlock

2019-02-13 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1097:
-

 Summary: Add genesis benchmark for BlockManager#allocateBlock
 Key: HDDS-1097
 URL: https://issues.apache.org/jira/browse/HDDS-1097
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


This Jira aims to add a genesis benchmark test for BlockManager#allocateBlock.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1081) Closing state containers should not be added to pipeline on SCM restart

2019-02-11 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1081:
-

 Summary: Closing state containers should not be added to pipeline 
on SCM restart
 Key: HDDS-1081
 URL: https://issues.apache.org/jira/browse/HDDS-1081
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


Currently while loading containers from db during scm start, containers in 
CLOSING state are added to pipeline. This can lead to error as the pipeline may 
have been removed from SCM db in such cases.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1047) Fix TestRatisPipelineProvider#testCreatePipelineWithFactor

2019-02-04 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1047:
-

 Summary: Fix TestRatisPipelineProvider#testCreatePipelineWithFactor
 Key: HDDS-1047
 URL: https://issues.apache.org/jira/browse/HDDS-1047
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-1045) Fix TestCloseContainerByPipeline#testIfCloseContainerCommandHandlerIsInvoked

2019-02-04 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-1045.
---
Resolution: Duplicate

> Fix TestCloseContainerByPipeline#testIfCloseContainerCommandHandlerIsInvoked
> 
>
> Key: HDDS-1045
> URL: https://issues.apache.org/jira/browse/HDDS-1045
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>Reporter: Lokesh Jain
>Priority: Major
> Fix For: 0.4.0
>
>
> TestCloseContainerByPipeline#testIfCloseContainerCommandHandlerIsInvoked is 
> failing with stack overflow error.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1045) Fix TestCloseContainerByPipeline#testIfCloseContainerCommandHandlerIsInvoked

2019-02-04 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1045:
-

 Summary: Fix 
TestCloseContainerByPipeline#testIfCloseContainerCommandHandlerIsInvoked
 Key: HDDS-1045
 URL: https://issues.apache.org/jira/browse/HDDS-1045
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
 Fix For: 0.4.0


TestCloseContainerByPipeline#testIfCloseContainerCommandHandlerIsInvoked is 
failing with stack overflow error.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1028) Improve logging in SCMPipelineManager

2019-01-29 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1028:
-

 Summary: Improve logging in SCMPipelineManager
 Key: HDDS-1028
 URL: https://issues.apache.org/jira/browse/HDDS-1028
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
 Fix For: 0.4.0


Currently SCMPipelineManager does not log events like pipeline creation and 
deletion. It would be a good idea to log such events.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1004) SCMContainerManager#updateContainerStateInternal fails for QUASI_CLOSE and FORCE_CLOSE events

2019-01-24 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-1004:
-

 Summary: SCMContainerManager#updateContainerStateInternal fails 
for QUASI_CLOSE and FORCE_CLOSE events
 Key: HDDS-1004
 URL: https://issues.apache.org/jira/browse/HDDS-1004
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


SCMContainerManager#updateContainerStateInternal currently fails for 
QUASI_CLOSE and FORCE_CLOSE events.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-982) Fix TestContainerDataYaml#testIncorrectContainerFile

2019-01-17 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-982:


 Summary: Fix TestContainerDataYaml#testIncorrectContainerFile
 Key: HDDS-982
 URL: https://issues.apache.org/jira/browse/HDDS-982
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-981) Block allocation should involve pipeline selection and then container selection

2019-01-16 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-981:


 Summary: Block allocation should involve pipeline selection and 
then container selection
 Key: HDDS-981
 URL: https://issues.apache.org/jira/browse/HDDS-981
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently SCM maintains a list of preallocated containers and allocates blocks 
from these containers. This approach does not work well with dynamics of the 
cluster where new nodes are being added and pipelines are destroyed. New 
containers are not created until all the preallocated containers are exhausted.

The Jira aims to establish a criteria in block allocation where first a 
pipeline is selected amongst the available pipelines and then a container is 
selected in that pipeline. In order to handle the dynamics of the cluster a 
fixed interval pipeline creator job can be launched which creates pipelines in 
the system.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-971) ContainerDataConstructor throws exception on QUASI_CLOSED and UNHEALTHY container state

2019-01-10 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-971:


 Summary: ContainerDataConstructor throws exception on QUASI_CLOSED 
and UNHEALTHY container state
 Key: HDDS-971
 URL: https://issues.apache.org/jira/browse/HDDS-971
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


We need to define QUASI_CLOSED and UNHEALTHY state in ContainerDataConstructor 
class logic. Currently the code uses switch-case for determining the state. The 
logic can be replaced with 
{code:java}
ContainerProtos.ContainerDataProto.State.valueOf(state)
{code}
call. The Jira also fixes test failure TestKeys#testPutAndGetKeyWithDnRestart.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-959) KeyOutputStream should handle retry failures

2019-01-03 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-959:


 Summary: KeyOutputStream should handle retry failures
 Key: HDDS-959
 URL: https://issues.apache.org/jira/browse/HDDS-959
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


With ratis version updated to 0.4.0-a8c4ca0-SNAPSHOT, retry failures are fatal 
for a raft client. If an operation in raft client does not succeed after 
maximum number of retries(RaftRetryFailureException) all subsequent operations 
are failed with AlreadyClosedException. This jira aims to handle such 
exceptions. Since we maintain a cache for clients in XceiverClientManager, the 
corresponding client needs to be invalidated in the cache.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-912) Update ozone to latest ratis snapshot build (0.4.0-3b0be02-SNAPSHOT)

2018-12-10 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-912:


 Summary: Update ozone to latest ratis snapshot build 
(0.4.0-3b0be02-SNAPSHOT)
 Key: HDDS-912
 URL: https://issues.apache.org/jira/browse/HDDS-912
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


We can update ratis snapshot build in ozone to 0.4.0-3b0be02-SNAPSHOT.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-871) XceiverClientRatis#watchForCommit closes raft client at every call

2018-12-07 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-871?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-871.
--
Resolution: Duplicate

> XceiverClientRatis#watchForCommit closes raft client at every call
> --
>
> Key: HDDS-871
> URL: https://issues.apache.org/jira/browse/HDDS-871
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>Reporter: Lokesh Jain
>Assignee: Lokesh Jain
>Priority: Major
> Fix For: 0.4.0
>
>
> Currently XceiverClientRatis#watchForCommit closes raft client at every call. 
> We should cache the raft client and reuse it for future watchForCommit calls.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-886) Unnecessary buffer copy in HddsDispatcher#dispatch

2018-11-30 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-886:


 Summary: Unnecessary buffer copy in HddsDispatcher#dispatch
 Key: HDDS-886
 URL: https://issues.apache.org/jira/browse/HDDS-886
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


In HddsDispatcher#dispatch precondition not null check converts container 
command to a string object. This is done even for a write chunk command which 
means we copy the chunk data to a string.
{code:java}
// code placeholderpublic ContainerCommandResponseProto dispatch(
ContainerCommandRequestProto msg) {
  LOG.trace("Command {}, trace ID: {} ", msg.getCmdType().toString(),
  msg.getTraceID());
  Preconditions.checkNotNull(msg.toString());

{code}
The precondition needs to check only the msg.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-871) XceiverClientRatis#watchForCommit closes raft client at every call

2018-11-24 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-871:


 Summary: XceiverClientRatis#watchForCommit closes raft client at 
every call
 Key: HDDS-871
 URL: https://issues.apache.org/jira/browse/HDDS-871
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


Currently XceiverClientRatis#watchForCommit closes raft client at every call. 
We should cache the raft client and reuse it for future watchForCommit calls.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-869) Fix log message in XceiverClientRatis#sendCommandAsync

2018-11-22 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-869:


 Summary: Fix log message in XceiverClientRatis#sendCommandAsync
 Key: HDDS-869
 URL: https://issues.apache.org/jira/browse/HDDS-869
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


The log message in XceiverClientRatis#sendCommandAsync is wrong. We should not 
print data in case of write chunk request.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-844) Add logic for pipeline teardown after timeout

2018-11-15 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-844:


 Summary: Add logic for pipeline teardown after timeout
 Key: HDDS-844
 URL: https://issues.apache.org/jira/browse/HDDS-844
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Affects Versions: 0.4.0
Reporter: Lokesh Jain
Assignee: Lokesh Jain


On receiving pipeline action we close the pipeline and wait for all containers 
to get closed. Currently pipeline is destroyed on datanodes only after all the 
containers have been closed. There is a possibility for containers to never get 
to CLOSED state if there is a two node failure. In such scenarios the pipeline 
needs to be destroyed and removed from SCM after a timeout.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-734) Remove create container logic from OzoneClient

2018-11-12 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-734?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-734.
--
Resolution: Duplicate

This issue has been fixed via HDDS-733.

> Remove create container logic from OzoneClient
> --
>
> Key: HDDS-734
> URL: https://issues.apache.org/jira/browse/HDDS-734
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: Ozone Client
>Reporter: Nanda kumar
>Assignee: Shashikant Banerjee
>Priority: Major
>
> After HDDS-733, the container will be created as part of the first chunk 
> write, we don't need explicit container creation code in {{OzoneClient}} 
> anymore.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-735) Remove ALLOCATED and CREATING state from ContainerStateManager

2018-11-12 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-735.
--
Resolution: Duplicate

This issue has been fixed via HDDS-733.

> Remove ALLOCATED and CREATING state from ContainerStateManager
> --
>
> Key: HDDS-735
> URL: https://issues.apache.org/jira/browse/HDDS-735
> Project: Hadoop Distributed Data Store
>  Issue Type: Improvement
>  Components: SCM
>Reporter: Nanda kumar
>Assignee: Lokesh Jain
>Priority: Major
>
> After HDDS-733 and HDDS-734, we don't need ALLOCATED and CREATING state for 
> containers in SCM. The container will move to OPEN state as soon as it is 
> allocated in SCM. Since the container creation happens as part of the first 
> chunk write and container creation operation in datanode idempotent we don't 
> have to worry about giving out the same container to multiple clients as soon 
> as it is allocated.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-802) Container State Manager should get open pipelines for allocating container

2018-11-05 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-802:


 Summary: Container State Manager should get open pipelines for 
allocating container
 Key: HDDS-802
 URL: https://issues.apache.org/jira/browse/HDDS-802
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.4.0


ContainerStateManager#allocateContainer currenlty calls getPipelines(type, 
factor) which returns pipelines of all states. This Jira aims to add another 
api getPipelines(type, factor, state)  which can be called by container state 
manager to get only the open pipelines.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-771) ChunkGroupOutputStream stream entries need to be properly updated on closed container exception

2018-10-31 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-771:


 Summary: ChunkGroupOutputStream stream entries need to be properly 
updated on closed container exception
 Key: HDDS-771
 URL: https://issues.apache.org/jira/browse/HDDS-771
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently ChunkGroupOutputStream does not increment the currentStreamIndex when 
a chunk write completes but there is no data in the buffer. This leads to 
overwriting of stream entry.

We also need to update the bcsid in case of closed container exception. The 
stream entry's bcsid needs to be updated with the bcsid of the committed block.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-744) Fix ASF license warning in PipelineNotFoundException class

2018-10-27 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-744:


 Summary: Fix ASF license warning in PipelineNotFoundException class
 Key: HDDS-744
 URL: https://issues.apache.org/jira/browse/HDDS-744
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-422) ContainerStateMachine.readStateMachineData throws OverlappingFileLockException

2018-10-23 Thread Lokesh Jain (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-422?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-422.
--
Resolution: Not A Bug

> ContainerStateMachine.readStateMachineData throws OverlappingFileLockException
> --
>
> Key: HDDS-422
> URL: https://issues.apache.org/jira/browse/HDDS-422
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: Ozone Datanode
>Reporter: Lokesh Jain
>Assignee: Lokesh Jain
>Priority: Major
>  Labels: Arches-Deferral-Candidate
>
>  
> {code:java}
> 2018-09-06 23:11:41,386 ERROR org.apache.ratis.server.impl.LogAppender: 
> GRpcLogAppender(d95c60fd-0e23-4237-8135-e05a326b952d_9858 -> 
> 954e7a3b-b20e-43a5-8f82-4381872aa7bb_9858) hit IOException while loadin
> g raft log
> org.apache.ratis.server.storage.RaftLogIOException: 
> d95c60fd-0e23-4237-8135-e05a326b952d_9858: Failed readStateMachineData for 
> (t:39, i:667)SMLOGENTRY, client-CD988394E416, cid=90
> at 
> org.apache.ratis.server.storage.RaftLog$EntryWithData.getEntry(RaftLog.java:360)
> at 
> org.apache.ratis.server.impl.LogAppender$LogEntryBuffer.getAppendRequest(LogAppender.java:165)
> at 
> org.apache.ratis.server.impl.LogAppender.createRequest(LogAppender.java:214)
> at 
> org.apache.ratis.grpc.server.GRpcLogAppender.appendLog(GRpcLogAppender.java:148)
> at 
> org.apache.ratis.grpc.server.GRpcLogAppender.runAppenderImpl(GRpcLogAppender.java:92)
> at org.apache.ratis.server.impl.LogAppender.runAppender(LogAppender.java:101)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.nio.channels.OverlappingFileLockException
> at sun.nio.ch.SharedFileLockTable.checkList(FileLockTable.java:255)
> at sun.nio.ch.SharedFileLockTable.add(FileLockTable.java:152)
> at 
> sun.nio.ch.AsynchronousFileChannelImpl.addToFileLockTable(AsynchronousFileChannelImpl.java:178)
> at 
> sun.nio.ch.SimpleAsynchronousFileChannelImpl.implLock(SimpleAsynchronousFileChannelImpl.java:185)
> at 
> sun.nio.ch.AsynchronousFileChannelImpl.lock(AsynchronousFileChannelImpl.java:118)
> at 
> org.apache.hadoop.ozone.container.keyvalue.helpers.ChunkUtils.readData(ChunkUtils.java:176)
> at 
> org.apache.hadoop.ozone.container.keyvalue.impl.ChunkManagerImpl.readChunk(ChunkManagerImpl.java:161)
> at 
> org.apache.hadoop.ozone.container.keyvalue.KeyValueHandler.handleReadChunk(KeyValueHandler.java:598)
> at 
> org.apache.hadoop.ozone.container.keyvalue.KeyValueHandler.handle(KeyValueHandler.java:201)
> at 
> org.apache.hadoop.ozone.container.common.impl.HddsDispatcher.dispatch(HddsDispatcher.java:142)
> at 
> org.apache.hadoop.ozone.container.common.transport.server.ratis.ContainerStateMachine.dispatchCommand(ContainerStateMachine.java:217)
> at 
> org.apache.hadoop.ozone.container.common.transport.server.ratis.ContainerStateMachine.readStateMachineData(ContainerStateMachine.java:289)
> at 
> org.apache.hadoop.ozone.container.common.transport.server.ratis.ContainerStateMachine.lambda$readStateMachineData$3(ContainerStateMachine.java:359)
> at 
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1590)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> ... 1 more
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-694) Plugin new Pipeline management code in SCM

2018-10-18 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-694:


 Summary: Plugin new Pipeline management code in SCM
 Key: HDDS-694
 URL: https://issues.apache.org/jira/browse/HDDS-694
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain


This Jira aims to plugin new pipeline management code in SCM. It removes the 
old pipeline related classes as well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-668) Replica Manager should use replica with latest delete transactionID

2018-10-16 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-668:


 Summary: Replica Manager should use replica with latest delete 
transactionID
 Key: HDDS-668
 URL: https://issues.apache.org/jira/browse/HDDS-668
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently replica manager does not use delete trasactionID for choosing the 
replica which will be replicated. This Jira aims to store delete transactionID 
for each replica so that replica manager can choose replica with latest delete 
transactionID.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-656) Add logic for pipeline report and action processing in new pipeline code

2018-10-14 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-656:


 Summary: Add logic for pipeline report and action processing in 
new pipeline code
 Key: HDDS-656
 URL: https://issues.apache.org/jira/browse/HDDS-656
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain


As part of pipeline refactoring, new pipeline management classes were added as 
part of HDDS-587. This Jira adds logic for pipeline report and action 
processing in the new code.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-587) Add new classes for pipeline management

2018-10-08 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-587:


 Summary: Add new classes for pipeline management
 Key: HDDS-587
 URL: https://issues.apache.org/jira/browse/HDDS-587
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.3.0


This Jira adds new classes and corresponding unit tests for pipeline management 
in SCM. The old classes will be removed in a subsequent jira.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-561) Move Node2ContainerMap and Node2PipelineMap to NodeManager

2018-09-27 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-561:


 Summary: Move Node2ContainerMap and Node2PipelineMap to NodeManager
 Key: HDDS-561
 URL: https://issues.apache.org/jira/browse/HDDS-561
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-480) RaftLog should make sure appendEntries term are incremental in nature

2018-09-17 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-480:


 Summary: RaftLog should make sure appendEntries term are 
incremental in nature
 Key: HDDS-480
 URL: https://issues.apache.org/jira/browse/HDDS-480
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently raft log does not make sure that any appendEntry has a term greater 
than or equal to the last applied entry's term in the log. This Jira aims to 
add that check.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-475) Block Allocation returns same BlockID on different keys creation

2018-09-16 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-475:


 Summary: Block Allocation returns same BlockID on different keys 
creation
 Key: HDDS-475
 URL: https://issues.apache.org/jira/browse/HDDS-475
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Affects Versions: 0.2.1
Reporter: Lokesh Jain
Assignee: Lokesh Jain


BlockManagerImpl#allocateBlock returns same BlockID. This leads to different 
key creations getting the same blockId.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-464) Fix TestCloseContainerHandlingByClient

2018-09-14 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-464:


 Summary: Fix TestCloseContainerHandlingByClient
 Key: HDDS-464
 URL: https://issues.apache.org/jira/browse/HDDS-464
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain


testBlockWriteViaRatis and testMultiBlockWrites2 fail with NPE and 
AssertionError respectively.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-433) ContainerStateMachine#readStateMachineData should properly build LogEntryProto

2018-09-11 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-433:


 Summary: ContainerStateMachine#readStateMachineData should 
properly build LogEntryProto
 Key: HDDS-433
 URL: https://issues.apache.org/jira/browse/HDDS-433
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Affects Versions: 0.2.1
Reporter: Lokesh Jain
Assignee: Lokesh Jain


ContainerStateMachine#readStateMachineData returns LogEntryProto with index set 
to 0. This leads to exception in Ratis. The LogEntryProto to return should be 
built over the input LogEntryProto.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-422) ContainerStateMachine.readStateMachineData throws OverlappingFileLockException

2018-09-10 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-422:


 Summary: ContainerStateMachine.readStateMachineData throws 
OverlappingFileLockException
 Key: HDDS-422
 URL: https://issues.apache.org/jira/browse/HDDS-422
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


 
{code:java}
2018-09-06 23:11:41,386 ERROR org.apache.ratis.server.impl.LogAppender: 
GRpcLogAppender(d95c60fd-0e23-4237-8135-e05a326b952d_9858 -> 
954e7a3b-b20e-43a5-8f82-4381872aa7bb_9858) hit IOException while loadin
g raft log
org.apache.ratis.server.storage.RaftLogIOException: 
d95c60fd-0e23-4237-8135-e05a326b952d_9858: Failed readStateMachineData for 
(t:39, i:667)SMLOGENTRY, client-CD988394E416, cid=90
at 
org.apache.ratis.server.storage.RaftLog$EntryWithData.getEntry(RaftLog.java:360)
at 
org.apache.ratis.server.impl.LogAppender$LogEntryBuffer.getAppendRequest(LogAppender.java:165)
at org.apache.ratis.server.impl.LogAppender.createRequest(LogAppender.java:214)
at 
org.apache.ratis.grpc.server.GRpcLogAppender.appendLog(GRpcLogAppender.java:148)
at 
org.apache.ratis.grpc.server.GRpcLogAppender.runAppenderImpl(GRpcLogAppender.java:92)
at org.apache.ratis.server.impl.LogAppender.runAppender(LogAppender.java:101)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.nio.channels.OverlappingFileLockException
at sun.nio.ch.SharedFileLockTable.checkList(FileLockTable.java:255)
at sun.nio.ch.SharedFileLockTable.add(FileLockTable.java:152)
at 
sun.nio.ch.AsynchronousFileChannelImpl.addToFileLockTable(AsynchronousFileChannelImpl.java:178)
at 
sun.nio.ch.SimpleAsynchronousFileChannelImpl.implLock(SimpleAsynchronousFileChannelImpl.java:185)
at 
sun.nio.ch.AsynchronousFileChannelImpl.lock(AsynchronousFileChannelImpl.java:118)
at 
org.apache.hadoop.ozone.container.keyvalue.helpers.ChunkUtils.readData(ChunkUtils.java:176)
at 
org.apache.hadoop.ozone.container.keyvalue.impl.ChunkManagerImpl.readChunk(ChunkManagerImpl.java:161)
at 
org.apache.hadoop.ozone.container.keyvalue.KeyValueHandler.handleReadChunk(KeyValueHandler.java:598)
at 
org.apache.hadoop.ozone.container.keyvalue.KeyValueHandler.handle(KeyValueHandler.java:201)
at 
org.apache.hadoop.ozone.container.common.impl.HddsDispatcher.dispatch(HddsDispatcher.java:142)
at 
org.apache.hadoop.ozone.container.common.transport.server.ratis.ContainerStateMachine.dispatchCommand(ContainerStateMachine.java:217)
at 
org.apache.hadoop.ozone.container.common.transport.server.ratis.ContainerStateMachine.readStateMachineData(ContainerStateMachine.java:289)
at 
org.apache.hadoop.ozone.container.common.transport.server.ratis.ContainerStateMachine.lambda$readStateMachineData$3(ContainerStateMachine.java:359)
at 
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1590)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
... 1 more
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-416) Fix bug in ChunkInputStreamEntry

2018-09-07 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-416:


 Summary: Fix bug in ChunkInputStreamEntry
 Key: HDDS-416
 URL: https://issues.apache.org/jira/browse/HDDS-416
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Client
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


ChunkInputStreamEntry maintains currentPosition field. This field is redundant 
and can be replaced by getPos().



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-397) Handle deletion for keys with no blocks

2018-09-03 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-397:


 Summary: Handle deletion for keys with no blocks
 Key: HDDS-397
 URL: https://issues.apache.org/jira/browse/HDDS-397
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Manager
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


Keys which do not contain blocks can be deleted directly from OzoneManager.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-386) Create a datanode cli

2018-08-30 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-386:


 Summary: Create a datanode cli
 Key: HDDS-386
 URL: https://issues.apache.org/jira/browse/HDDS-386
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


For block deletion we need a debug cli on the datanode to know the state of the 
containers and number of chunks present in the container.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-325) Add event watcher for delete blocks command

2018-08-05 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-325:


 Summary: Add event watcher for delete blocks command
 Key: HDDS-325
 URL: https://issues.apache.org/jira/browse/HDDS-325
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode, SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain


This Jira aims to add watcher for deleteBlocks command. It removes the current 
rpc call required for datanode to send the acknowledgement for deleteBlocks.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-308) SCM should identify a container with pending deletes using container reports

2018-07-31 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-308:


 Summary: SCM should identify a container with pending deletes 
using container reports
 Key: HDDS-308
 URL: https://issues.apache.org/jira/browse/HDDS-308
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


SCM should fire an event when it finds using container report that a 
container's deleteTransactionID does not match SCM's deleteTransactionId.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-13765) Fix javadoc for FSDirMkdirOp#createParentDirectories

2018-07-25 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDFS-13765:
--

 Summary: Fix javadoc for FSDirMkdirOp#createParentDirectories
 Key: HDFS-13765
 URL: https://issues.apache.org/jira/browse/HDFS-13765
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Javadoc needs to be fixed for FSDirMkdirOp#createParentDirectories.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-279) DeleteBlocks command should not be sent for open containers

2018-07-21 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-279:


 Summary: DeleteBlocks command should not be sent for open 
containers
 Key: HDDS-279
 URL: https://issues.apache.org/jira/browse/HDDS-279
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


Currently SCM sends delete blocks command for open containers as well. This 
Jira aims to disable it.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-273) DeleteLog entries should be purged only after corresponding dns commit the transaction

2018-07-20 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-273:


 Summary: DeleteLog entries should be purged only after 
corresponding dns commit the transaction
 Key: HDDS-273
 URL: https://issues.apache.org/jira/browse/HDDS-273
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1
 Attachments: HDDS-273.001.patch

Currently DeleteBlockLogImpl removes an entry corresponding to a deleteBlock as 
soon as it receives an acknowledgement from a single DN. The entries should be 
deleted only after all corresponding dns give an acknowledgement for the delete 
command.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-254) Fix TestStorageContainerManager#testBlockDeletingThrottling

2018-07-13 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-254:


 Summary: Fix 
TestStorageContainerManager#testBlockDeletingThrottling
 Key: HDDS-254
 URL: https://issues.apache.org/jira/browse/HDDS-254
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-253) SCMBlockDeletingService should publish events for delete blocks

2018-07-13 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-253:


 Summary: SCMBlockDeletingService should publish events for delete 
blocks
 Key: HDDS-253
 URL: https://issues.apache.org/jira/browse/HDDS-253
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: SCM
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


SCMBlockDeletingService should publish events for delete Blocks command. 
Currently it directly makes a call to SCMNodeManager.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-251) Integrate BlockDeletingService in KeyValueHandler

2018-07-11 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-251:


 Summary: Integrate BlockDeletingService in KeyValueHandler
 Key: HDDS-251
 URL: https://issues.apache.org/jira/browse/HDDS-251
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


This Jira aims to integrate BlockDeletingService in KeyValueHandler. It also 
fixes the unit tests related to delete blocks.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-178) DeleteBlocks should not be handled by open containers

2018-06-19 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-178:


 Summary: DeleteBlocks should not be handled by open containers
 Key: HDDS-178
 URL: https://issues.apache.org/jira/browse/HDDS-178
 Project: Hadoop Distributed Data Store
  Issue Type: Task
  Components: Ozone Datanode
Reporter: Lokesh Jain
Assignee: Lokesh Jain


In the case of open containers deleteBlocks command just adds an entry in the 
log but does not delete the blocks. These blocks are deleted only when 
container is closed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-159) RestClient: Implement list operations for volume, bucket and keys

2018-06-10 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-159:


 Summary: RestClient: Implement list operations for volume, bucket 
and keys
 Key: HDDS-159
 URL: https://issues.apache.org/jira/browse/HDDS-159
 Project: Hadoop Distributed Data Store
  Issue Type: Task
  Components: Ozone Client
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1
 Attachments: HDDS-159.001.patch

Currently RestClient does not provide implementation for list volume, list 
buckets and list keys. This Jira aims to add the implementation and necessary 
tests for the same.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-113) Rest and Rpc Client should verify resource name using HddsClientUtils

2018-05-23 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-113:


 Summary: Rest and Rpc Client should verify resource name using 
HddsClientUtils
 Key: HDDS-113
 URL: https://issues.apache.org/jira/browse/HDDS-113
 Project: Hadoop Distributed Data Store
  Issue Type: Task
  Components: Ozone Client
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


Currently Rest and Rpc client do not verify resource name. They should do so 
using HddsClientUtils#verifyResourceName.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-112) OzoneShell should support commands with url without scheme

2018-05-23 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-112:


 Summary: OzoneShell should support commands with url without scheme
 Key: HDDS-112
 URL: https://issues.apache.org/jira/browse/HDDS-112
 Project: Hadoop Distributed Data Store
  Issue Type: Task
  Components: Ozone Client
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


Currently urls need to have o3 or http scheme in OzoneShell. This Jira aims to 
support urls without scheme in OzoneShell. For such cases default o3 scheme 
would be used.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-109) Add reconnect logic for XceiverClientGrpc

2018-05-23 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-109:


 Summary: Add reconnect logic for XceiverClientGrpc
 Key: HDDS-109
 URL: https://issues.apache.org/jira/browse/HDDS-109
 Project: Hadoop Distributed Data Store
  Issue Type: Task
  Components: Ozone Client
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


We need to add reconnect logic in XceiverClientGrpc which allows it to 
reconnect in case of DN restart.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-73) Add acceptance tests for Ozone Shell

2018-05-15 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-73:
---

 Summary: Add acceptance tests for Ozone Shell
 Key: HDDS-73
 URL: https://issues.apache.org/jira/browse/HDDS-73
 Project: Hadoop Distributed Data Store
  Issue Type: Test
Reporter: Lokesh Jain
Assignee: Lokesh Jain


This Jira aims to add acceptance tests related to http, o3 scheme and various 
server port combinations in shell commands.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-72) Add numDeletedBlocks field in ContainerInfo

2018-05-15 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-72:
---

 Summary: Add numDeletedBlocks field in ContainerInfo
 Key: HDDS-72
 URL: https://issues.apache.org/jira/browse/HDDS-72
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain


This Jira aims to add numDeletedBlocks field in SCMContainerInfo and 
ContainerInfo in hdds.proto and StorageContainerDatanodeProtocol.proto 
respectively. This field would serve as a generation Id for block deletes as it 
is monotonically increasing in nature and would also give information about the 
number of deleted blocks.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-47) Add acceptance tests for Ozone Shell

2018-05-11 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-47:
---

 Summary: Add acceptance tests for Ozone Shell
 Key: HDDS-47
 URL: https://issues.apache.org/jira/browse/HDDS-47
 Project: Hadoop Distributed Data Store
  Issue Type: Test
Reporter: Lokesh Jain
Assignee: Lokesh Jain






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-36) Ozone: Enable checkstyle for HDDS

2018-05-10 Thread Lokesh Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDDS-36?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lokesh Jain resolved HDDS-36.
-
Resolution: Not A Problem

> Ozone: Enable checkstyle for HDDS
> -
>
> Key: HDDS-36
> URL: https://issues.apache.org/jira/browse/HDDS-36
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>Reporter: Lokesh Jain
>Assignee: Lokesh Jain
>Priority: Major
> Fix For: 0.2.1
>
>
> Before merge we had disabled the checkstyle for hadoop-hdds and hadoop-ozone. 
> Since we have our own subprojects now we can enable them with hdds profile.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-36) Ozone: Enable checkstyle for HDDS

2018-05-09 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-36:
---

 Summary: Ozone: Enable checkstyle for HDDS
 Key: HDDS-36
 URL: https://issues.apache.org/jira/browse/HDDS-36
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Fix For: 0.2.1


Before merge we had disabled the checkstyle for hadoop-hdds and hadoop-ozone. 
Since we have our own subprojects now we can enable them with hdds profile.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-24) Ozone: Rename directory in ozonefs should be atomic

2018-05-07 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDDS-24:
---

 Summary: Ozone: Rename directory in ozonefs should be atomic
 Key: HDDS-24
 URL: https://issues.apache.org/jira/browse/HDDS-24
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
Reporter: Lokesh Jain
Assignee: Lokesh Jain


Currently rename in ozonefs is not atomic. While rename takes place another 
client might be adding a new file into the directory. Further if rename fails 
midway the directory will be in an inconsistent state.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-13458) Ozone: Fix javadoc issues in HDFS-7240

2018-04-16 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDFS-13458:
--

 Summary: Ozone: Fix javadoc issues in HDFS-7240
 Key: HDFS-13458
 URL: https://issues.apache.org/jira/browse/HDFS-13458
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Lokesh Jain






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-13456) Ozone: Update ozone to latest ratis snapshot build (0.1.1-alpha-4309324-SNAPSHOT)

2018-04-16 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDFS-13456:
--

 Summary: Ozone: Update ozone to latest ratis snapshot build 
(0.1.1-alpha-4309324-SNAPSHOT)
 Key: HDFS-13456
 URL: https://issues.apache.org/jira/browse/HDFS-13456
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Lokesh Jain
Assignee: Lokesh Jain






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-13444) Ozone: Fix checkstyle issues in HDFS-7240

2018-04-13 Thread Lokesh Jain (JIRA)
Lokesh Jain created HDFS-13444:
--

 Summary: Ozone: Fix checkstyle issues in HDFS-7240
 Key: HDFS-13444
 URL: https://issues.apache.org/jira/browse/HDFS-13444
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Lokesh Jain
Assignee: Lokesh Jain
 Attachments: HDFS-7240.007.patch





--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



  1   2   >