[
https://issues.apache.org/jira/browse/HDDS-8072?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Jyotirmoy Sinha updated HDDS-8072:
----------------------------------
Description:
After creating 13525 snapshots on a cluster across various volume/buckets there
is an OM error due to 'Error during Snapshot sst filtering'
ozone-om.log stacktrace -
{code:java}
2023-03-02 20:24:13,850 ERROR org.apache.hadoop.ozone.om.SstFilteringService:
Error during Snapshot sst filtering
java.io.IOException: Failed init RocksDB, db path :
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982,
exception :org.rocksdb.RocksDBException While open a file for random read:
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
Too many open files
at org.apache.hadoop.hdds.utils.db.RDBStore.<init>(RDBStore.java:180)
at
org.apache.hadoop.hdds.utils.db.DBStoreBuilder.build(DBStoreBuilder.java:219)
at
org.apache.hadoop.ozone.om.OmMetadataManagerImpl.loadDB(OmMetadataManagerImpl.java:481)
at
org.apache.hadoop.ozone.om.SstFilteringService$SstFilteringTask.call(SstFilteringService.java:147)
at
org.apache.hadoop.hdds.utils.BackgroundService$PeriodicalTask.lambda$run$0(BackgroundService.java:121)
at
java.util.concurrent.CompletableFuture$AsyncRun.run(CompletableFuture.java:1640)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: class
org.apache.hadoop.hdds.utils.db.RocksDatabase: Failed to open
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982;
status : IOError; message : While open a file for random read:
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
Too many open files
at
org.apache.hadoop.hdds.utils.HddsServerUtil.toIOException(HddsServerUtil.java:576)
at
org.apache.hadoop.hdds.utils.db.RocksDatabase.toIOException(RocksDatabase.java:85)
at
org.apache.hadoop.hdds.utils.db.RocksDatabase.open(RocksDatabase.java:162)
at org.apache.hadoop.hdds.utils.db.RDBStore.<init>(RDBStore.java:116)
... 12 more
Caused by: org.rocksdb.RocksDBException: While open a file for random read:
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
Too many open files
at org.rocksdb.RocksDB.openROnly(Native Method)
at org.rocksdb.RocksDB.openReadOnly(RocksDB.java:488)
at org.rocksdb.RocksDB.openReadOnly(RocksDB.java:443)
at
org.apache.hadoop.hdds.utils.db.managed.ManagedRocksDB.openReadOnly(ManagedRocksDB.java:45)
at
org.apache.hadoop.hdds.utils.db.RocksDatabase.open(RocksDatabase.java:145)
... 13 more {code}
was:
After creating 13525 snapshot on a cluster across various volume/buckets there
is an OM error due to 'Error during Snapshot sst filtering'
ozone-om.log stacktrace -
{code:java}
2023-03-02 20:24:13,850 ERROR org.apache.hadoop.ozone.om.SstFilteringService:
Error during Snapshot sst filtering
java.io.IOException: Failed init RocksDB, db path :
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982,
exception :org.rocksdb.RocksDBException While open a file for random read:
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
Too many open files
at org.apache.hadoop.hdds.utils.db.RDBStore.<init>(RDBStore.java:180)
at
org.apache.hadoop.hdds.utils.db.DBStoreBuilder.build(DBStoreBuilder.java:219)
at
org.apache.hadoop.ozone.om.OmMetadataManagerImpl.loadDB(OmMetadataManagerImpl.java:481)
at
org.apache.hadoop.ozone.om.SstFilteringService$SstFilteringTask.call(SstFilteringService.java:147)
at
org.apache.hadoop.hdds.utils.BackgroundService$PeriodicalTask.lambda$run$0(BackgroundService.java:121)
at
java.util.concurrent.CompletableFuture$AsyncRun.run(CompletableFuture.java:1640)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: class
org.apache.hadoop.hdds.utils.db.RocksDatabase: Failed to open
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982;
status : IOError; message : While open a file for random read:
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
Too many open files
at
org.apache.hadoop.hdds.utils.HddsServerUtil.toIOException(HddsServerUtil.java:576)
at
org.apache.hadoop.hdds.utils.db.RocksDatabase.toIOException(RocksDatabase.java:85)
at
org.apache.hadoop.hdds.utils.db.RocksDatabase.open(RocksDatabase.java:162)
at org.apache.hadoop.hdds.utils.db.RDBStore.<init>(RDBStore.java:116)
... 12 more
Caused by: org.rocksdb.RocksDBException: While open a file for random read:
/var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
Too many open files
at org.rocksdb.RocksDB.openROnly(Native Method)
at org.rocksdb.RocksDB.openReadOnly(RocksDB.java:488)
at org.rocksdb.RocksDB.openReadOnly(RocksDB.java:443)
at
org.apache.hadoop.hdds.utils.db.managed.ManagedRocksDB.openReadOnly(ManagedRocksDB.java:45)
at
org.apache.hadoop.hdds.utils.db.RocksDatabase.open(RocksDatabase.java:145)
... 13 more {code}
> [snapshot] OM error due to 'Error during Snapshot sst filtering'
> ----------------------------------------------------------------
>
> Key: HDDS-8072
> URL: https://issues.apache.org/jira/browse/HDDS-8072
> Project: Apache Ozone
> Issue Type: Bug
> Components: Ozone Manager
> Reporter: Jyotirmoy Sinha
> Priority: Major
> Labels: ozone-snapshot
>
> After creating 13525 snapshots on a cluster across various volume/buckets
> there is an OM error due to 'Error during Snapshot sst filtering'
> ozone-om.log stacktrace -
> {code:java}
> 2023-03-02 20:24:13,850 ERROR org.apache.hadoop.ozone.om.SstFilteringService:
> Error during Snapshot sst filtering
> java.io.IOException: Failed init RocksDB, db path :
> /var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982,
> exception :org.rocksdb.RocksDBException While open a file for random read:
> /var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
> Too many open files
> at org.apache.hadoop.hdds.utils.db.RDBStore.<init>(RDBStore.java:180)
> at
> org.apache.hadoop.hdds.utils.db.DBStoreBuilder.build(DBStoreBuilder.java:219)
> at
> org.apache.hadoop.ozone.om.OmMetadataManagerImpl.loadDB(OmMetadataManagerImpl.java:481)
> at
> org.apache.hadoop.ozone.om.SstFilteringService$SstFilteringTask.call(SstFilteringService.java:147)
> at
> org.apache.hadoop.hdds.utils.BackgroundService$PeriodicalTask.lambda$run$0(BackgroundService.java:121)
> at
> java.util.concurrent.CompletableFuture$AsyncRun.run(CompletableFuture.java:1640)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
> at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: java.io.IOException: class
> org.apache.hadoop.hdds.utils.db.RocksDatabase: Failed to open
> /var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982;
> status : IOError; message : While open a file for random read:
> /var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
> Too many open files
> at
> org.apache.hadoop.hdds.utils.HddsServerUtil.toIOException(HddsServerUtil.java:576)
> at
> org.apache.hadoop.hdds.utils.db.RocksDatabase.toIOException(RocksDatabase.java:85)
> at
> org.apache.hadoop.hdds.utils.db.RocksDatabase.open(RocksDatabase.java:162)
> at org.apache.hadoop.hdds.utils.db.RDBStore.<init>(RDBStore.java:116)
> ... 12 more
> Caused by: org.rocksdb.RocksDBException: While open a file for random read:
> /var/lib/hadoop-ozone/om/data/db.snapshots/om.db-efd0f99b-3689-4ed6-a671-d2cb4d5fa982/000152.sst:
> Too many open files
> at org.rocksdb.RocksDB.openROnly(Native Method)
> at org.rocksdb.RocksDB.openReadOnly(RocksDB.java:488)
> at org.rocksdb.RocksDB.openReadOnly(RocksDB.java:443)
> at
> org.apache.hadoop.hdds.utils.db.managed.ManagedRocksDB.openReadOnly(ManagedRocksDB.java:45)
> at
> org.apache.hadoop.hdds.utils.db.RocksDatabase.open(RocksDatabase.java:145)
> ... 13 more {code}
>
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]