[
https://issues.apache.org/jira/browse/FLINK-8845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16397393#comment-16397393
]
ASF GitHub Bot commented on FLINK-8845:
---------------------------------------
Github user bowenli86 commented on a diff in the pull request:
https://github.com/apache/flink/pull/5650#discussion_r174231142
--- Diff:
flink-state-backends/flink-statebackend-rocksdb/src/test/java/org/apache/flink/contrib/streaming/state/RocksDBWriteBatchWrapperTest.java
---
@@ -0,0 +1,159 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.contrib.streaming.state;
+
+import org.apache.flink.api.java.tuple.Tuple2;
+
+import org.junit.Assert;
+import org.junit.Ignore;
+import org.junit.Rule;
+import org.junit.Test;
+import org.junit.rules.TemporaryFolder;
+import org.rocksdb.ColumnFamilyDescriptor;
+import org.rocksdb.ColumnFamilyHandle;
+import org.rocksdb.RocksDB;
+import org.rocksdb.WriteOptions;
+
+import java.util.ArrayList;
+import java.util.List;
+
+/**
+ * Tests to guard {@link RocksDBWriteBatchWrapper}.
+ */
+public class RocksDBWriteBatchWrapperTest {
+
+ @Rule
+ public TemporaryFolder folder = new TemporaryFolder();
+
+ @Test
+ public void basicTest() throws Exception {
+
+ List<Tuple2<byte[], byte[]>> data = new ArrayList<>(10000);
+ for (int i = 0; i < 10000; ++i) {
+ data.add(new Tuple2<>(("key:" + i).getBytes(),
("value:" + i).getBytes()));
+ }
+
+ try (RocksDB db =
RocksDB.open(folder.newFolder().getAbsolutePath());
+ WriteOptions options = new
WriteOptions().setDisableWAL(true);
+ ColumnFamilyHandle handle = db.createColumnFamily(new
ColumnFamilyDescriptor("test".getBytes()));
+ RocksDBWriteBatchWrapper writeBatchWrapper = new
RocksDBWriteBatchWrapper(db, options, 200)) {
+
+ // insert data
+ for (Tuple2<byte[], byte[]> item : data) {
+ writeBatchWrapper.put(handle, item.f0, item.f1);
+ }
+ writeBatchWrapper.flush();
+
+ // valid result
+ for (Tuple2<byte[], byte[]> item : data) {
+ Assert.assertArrayEquals(item.f1,
db.get(handle, item.f0));
+ }
+ }
+ }
+
+ @Test
+ @Ignore
+ public void benchMark() throws Exception {
+
+ // put with disableWAL=true VS put with disableWAL=false
+ System.out.println("--> put with disableWAL=true VS put with
disableWAL=false <--");
+ benchMarkHelper(1_000, true, WRITETYPE.PUT);
+ benchMarkHelper(1_000, false, WRITETYPE.PUT);
+
+ benchMarkHelper(10_000, true, WRITETYPE.PUT);
+ benchMarkHelper(10_000, false, WRITETYPE.PUT);
+
+ benchMarkHelper(100_000, true, WRITETYPE.PUT);
+ benchMarkHelper(100_000, false, WRITETYPE.PUT);
+
+ benchMarkHelper(1_000_000, true, WRITETYPE.PUT);
+ benchMarkHelper(1_000_000, false, WRITETYPE.PUT);
+
+ // put with disableWAL=true VS write batch with disableWAL=false
+ System.out.println("--> put with disableWAL=true VS write batch
with disableWAL=false <--");
+ benchMarkHelper(1_000, true, WRITETYPE.PUT);
+ benchMarkHelper(1_000, false, WRITETYPE.WRITE_BATCH);
+
+ benchMarkHelper(10_000, true, WRITETYPE.PUT);
+ benchMarkHelper(10_000, false, WRITETYPE.WRITE_BATCH);
+
+ benchMarkHelper(100_000, true, WRITETYPE.PUT);
+ benchMarkHelper(100_000, false, WRITETYPE.WRITE_BATCH);
+
+ benchMarkHelper(1_000_000, true, WRITETYPE.PUT);
+ benchMarkHelper(1_000_000, false, WRITETYPE.WRITE_BATCH);
+
+ // write batch with disableWAL=true VS write batch disableWAL =
true
+ System.out.println("--> write batch with disableWAL=true VS
write batch disableWAL = true <--");
+ benchMarkHelper(1_000, true, WRITETYPE.WRITE_BATCH);
+ benchMarkHelper(1_000, false, WRITETYPE.WRITE_BATCH);
+
+ benchMarkHelper(10_000, true, WRITETYPE.WRITE_BATCH);
+ benchMarkHelper(10_000, false, WRITETYPE.WRITE_BATCH);
+
+ benchMarkHelper(100_000, true, WRITETYPE.WRITE_BATCH);
+ benchMarkHelper(100_000, false, WRITETYPE.WRITE_BATCH);
+
+ benchMarkHelper(1_000_000, true, WRITETYPE.WRITE_BATCH);
+ benchMarkHelper(1_000_000, false, WRITETYPE.WRITE_BATCH);
+ }
+
+ private enum WRITETYPE {PUT, WRITE_BATCH}
+
+ private void benchMarkHelper(int number, boolean disableWAL, WRITETYPE
type) throws Exception {
+
+ List<Tuple2<byte[], byte[]>> data = new ArrayList<>(number);
+ for (int i = 0; i < number; ++i) {
+ data.add(new Tuple2<>(("key key key key key key" +
i).getBytes(),
+ ("value value value value value" +
i).getBytes()));
+ }
+
+ System.out.print("number:" + number);
+ switch (type) {
+ case PUT:
+ try (RocksDB db =
RocksDB.open(folder.newFolder().getAbsolutePath());
+ WriteOptions options = new
WriteOptions().setDisableWAL(disableWAL);
+ ColumnFamilyHandle handle =
db.createColumnFamily(new ColumnFamilyDescriptor("test".getBytes()))) {
+ long t1 = System.currentTimeMillis();
+ for (Tuple2<byte[], byte[]> item :
data) {
+ db.put(handle, options,
item.f0, item.f1);
+ }
+ long t2 = System.currentTimeMillis();
+ System.out.println(" put cost:" + (t2 -
t1) + " ms");
--- End diff --
ditto
> Use WriteBatch to improve performance for recovery in RocksDB backend
> ---------------------------------------------------------------------
>
> Key: FLINK-8845
> URL: https://issues.apache.org/jira/browse/FLINK-8845
> Project: Flink
> Issue Type: Improvement
> Components: State Backends, Checkpointing
> Affects Versions: 1.5.0
> Reporter: Sihua Zhou
> Assignee: Sihua Zhou
> Priority: Major
> Fix For: 1.6.0
>
>
> Base on {{WriteBatch}} we could get 30% ~ 50% performance lift when loading
> data into RocksDB.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)