Yingyi Bu has submitted this change and it was merged. Change subject: Remove unused methods in DatasetUtil, IndexUtil, and MetadataProvider. ......................................................................
Remove unused methods in DatasetUtil, IndexUtil, and MetadataProvider. Change-Id: I60352d052f684a65fd823a1c7aa33b794bd7f476 Reviewed-on: https://asterix-gerrit.ics.uci.edu/1741 Sonar-Qube: Jenkins <[email protected]> Reviewed-by: Till Westmann <[email protected]> Integration-Tests: Jenkins <[email protected]> Tested-by: Jenkins <[email protected]> --- M asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java M asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java M asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java 3 files changed, 1 insertion(+), 133 deletions(-) Approvals: Till Westmann: Looks good to me, approved Jenkins: Verified; No violations found; Verified Objections: Jenkins: Violations found diff --git a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java index 5934f5e..750f89b 100644 --- a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java +++ b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java @@ -35,7 +35,6 @@ import org.apache.asterix.common.dataflow.LSMTreeInsertDeleteOperatorDescriptor; import org.apache.asterix.common.exceptions.AsterixException; import org.apache.asterix.common.exceptions.CompilationException; -import org.apache.asterix.common.library.ILibraryManager; import org.apache.asterix.common.transactions.JobId; import org.apache.asterix.common.utils.StoragePathUtil; import org.apache.asterix.external.adapter.factory.LookupAdapterFactory; @@ -127,9 +126,7 @@ import org.apache.hyracks.dataflow.std.file.IFileSplitProvider; import org.apache.hyracks.dataflow.std.result.ResultWriterOperatorDescriptor; import org.apache.hyracks.storage.am.btree.dataflow.BTreeSearchOperatorDescriptor; -import org.apache.hyracks.storage.am.common.api.IMetadataPageManagerFactory; import org.apache.hyracks.storage.am.common.api.IModificationOperationCallbackFactory; -import org.apache.hyracks.storage.am.common.api.IPrimitiveValueProviderFactory; import org.apache.hyracks.storage.am.common.api.ISearchOperationCallbackFactory; import org.apache.hyracks.storage.am.common.dataflow.IIndexDataflowHelperFactory; import org.apache.hyracks.storage.am.common.dataflow.IndexDataflowHelperFactory; @@ -144,10 +141,7 @@ private final ICcApplicationContext appCtx; private final IStorageComponentProvider storaegComponentProvider; - private final IMetadataPageManagerFactory metadataPageManagerFactory; - private final IPrimitiveValueProviderFactory primitiveValueProviderFactory; private final StorageProperties storageProperties; - private final ILibraryManager libraryManager; private final Dataverse defaultDataverse; private final LockList locks; @@ -170,9 +164,6 @@ this.defaultDataverse = defaultDataverse; this.storaegComponentProvider = componentProvider; storageProperties = appCtx.getStorageProperties(); - libraryManager = appCtx.getLibraryManager(); - metadataPageManagerFactory = componentProvider.getMetadataPageManagerFactory(); - primitiveValueProviderFactory = componentProvider.getPrimitiveValueProviderFactory(); locks = new LockList(); } @@ -195,10 +186,6 @@ @Override public Map<String, String> getConfig() { return config; - } - - public ILibraryManager getLibraryManager() { - return libraryManager; } public void setJobId(JobId jobId) { @@ -259,10 +246,6 @@ public void setResultSerializerFactoryProvider(IResultSerializerFactoryProvider rafp) { this.resultSerializerFactoryProvider = rafp; - } - - public IResultSerializerFactoryProvider getResultSerializerFactoryProvider() { - return resultSerializerFactoryProvider; } public boolean isWriteTransaction() { @@ -379,10 +362,6 @@ } catch (AsterixException e) { throw new AlgebricksException(e); } - } - - public static AlgebricksAbsolutePartitionConstraint determineLocationConstraint(FeedDataSource feedDataSource) { - return new AlgebricksAbsolutePartitionConstraint(feedDataSource.getLocations()); } protected Pair<IOperatorDescriptor, AlgebricksPartitionConstraint> buildLoadableDatasetScan( diff --git a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java index e2cfb8b..098645e 100644 --- a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java +++ b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java @@ -53,9 +53,7 @@ import org.apache.hyracks.algebricks.common.exceptions.AlgebricksException; import org.apache.hyracks.algebricks.common.utils.Pair; import org.apache.hyracks.algebricks.data.IBinaryComparatorFactoryProvider; -import org.apache.hyracks.algebricks.data.IBinaryHashFunctionFactoryProvider; import org.apache.hyracks.api.dataflow.value.IBinaryComparatorFactory; -import org.apache.hyracks.api.dataflow.value.IBinaryHashFunctionFactory; import org.apache.hyracks.api.dataflow.value.ISerializerDeserializer; import org.apache.hyracks.api.dataflow.value.ITypeTraits; import org.apache.hyracks.api.exceptions.HyracksDataException; @@ -66,8 +64,8 @@ import org.apache.hyracks.storage.am.common.build.IndexBuilderFactory; import org.apache.hyracks.storage.am.common.dataflow.IIndexDataflowHelperFactory; import org.apache.hyracks.storage.am.common.dataflow.IndexCreateOperatorDescriptor; -import org.apache.hyracks.storage.am.common.dataflow.IndexDropOperatorDescriptor; import org.apache.hyracks.storage.am.common.dataflow.IndexDataflowHelperFactory; +import org.apache.hyracks.storage.am.common.dataflow.IndexDropOperatorDescriptor; import org.apache.hyracks.storage.am.lsm.common.api.ILSMMergePolicyFactory; import org.apache.hyracks.storage.am.lsm.common.dataflow.LSMTreeIndexCompactOperatorDescriptor; import org.apache.hyracks.storage.common.IResourceFactory; @@ -77,97 +75,9 @@ /* * Dataset related operations */ - public static final byte OP_READ = 0x00; - public static final byte OP_INSERT = 0x01; - public static final byte OP_DELETE = 0x02; public static final byte OP_UPSERT = 0x03; private DatasetUtil() { - } - - public static IBinaryComparatorFactory[] computeKeysBinaryComparatorFactories(Dataset dataset, ARecordType itemType, - ARecordType metaItemType, IBinaryComparatorFactoryProvider comparatorFactoryProvider) - throws AlgebricksException { - List<List<String>> partitioningKeys = dataset.getPrimaryKeys(); - IBinaryComparatorFactory[] bcfs = new IBinaryComparatorFactory[partitioningKeys.size()]; - if (dataset.getDatasetType() == DatasetType.EXTERNAL) { - // Get comparators for RID fields. - for (int i = 0; i < partitioningKeys.size(); i++) { - try { - bcfs[i] = IndexingConstants.getComparatorFactory(i); - } catch (AsterixException e) { - throw new AlgebricksException(e); - } - } - } else { - InternalDatasetDetails dsd = (InternalDatasetDetails) dataset.getDatasetDetails(); - for (int i = 0; i < partitioningKeys.size(); i++) { - IAType keyType = (dataset.hasMetaPart() && dsd.getKeySourceIndicator().get(i).intValue() == 1) - ? metaItemType.getSubFieldType(partitioningKeys.get(i)) - : itemType.getSubFieldType(partitioningKeys.get(i)); - bcfs[i] = comparatorFactoryProvider.getBinaryComparatorFactory(keyType, true); - } - } - return bcfs; - } - - public static int[] createBloomFilterKeyFields(Dataset dataset) throws AlgebricksException { - if (dataset.getDatasetType() == DatasetType.EXTERNAL) { - throw new AlgebricksException("not implemented"); - } - List<List<String>> partitioningKeys = dataset.getPrimaryKeys(); - int[] bloomFilterKeyFields = new int[partitioningKeys.size()]; - for (int i = 0; i < partitioningKeys.size(); ++i) { - bloomFilterKeyFields[i] = i; - } - return bloomFilterKeyFields; - } - - public static IBinaryHashFunctionFactory[] computeKeysBinaryHashFunFactories(Dataset dataset, ARecordType itemType, - IBinaryHashFunctionFactoryProvider hashFunProvider) throws AlgebricksException { - if (dataset.getDatasetType() == DatasetType.EXTERNAL) { - throw new AlgebricksException("not implemented"); - } - List<List<String>> partitioningKeys = dataset.getPrimaryKeys(); - IBinaryHashFunctionFactory[] bhffs = new IBinaryHashFunctionFactory[partitioningKeys.size()]; - for (int i = 0; i < partitioningKeys.size(); i++) { - IAType keyType = itemType.getSubFieldType(partitioningKeys.get(i)); - bhffs[i] = hashFunProvider.getBinaryHashFunctionFactory(keyType); - } - return bhffs; - } - - public static ITypeTraits[] computeTupleTypeTraits(Dataset dataset, ARecordType itemType, ARecordType metaItemType) - throws AlgebricksException { - if (dataset.getDatasetType() == DatasetType.EXTERNAL) { - throw new AlgebricksException("not implemented"); - } - List<List<String>> partitioningKeys = dataset.getPrimaryKeys(); - int numKeys = partitioningKeys.size(); - ITypeTraits[] typeTraits; - if (metaItemType != null) { - typeTraits = new ITypeTraits[numKeys + 2]; - List<Integer> indicator = ((InternalDatasetDetails) dataset.getDatasetDetails()).getKeySourceIndicator(); - typeTraits[numKeys + 1] = TypeTraitProvider.INSTANCE.getTypeTrait(metaItemType); - for (int i = 0; i < numKeys; i++) { - IAType keyType; - if (indicator.get(i) == 0) { - keyType = itemType.getSubFieldType(partitioningKeys.get(i)); - } else { - keyType = metaItemType.getSubFieldType(partitioningKeys.get(i)); - } - typeTraits[i] = TypeTraitProvider.INSTANCE.getTypeTrait(keyType); - } - } else { - typeTraits = new ITypeTraits[numKeys + 1]; - for (int i = 0; i < numKeys; i++) { - IAType keyType; - keyType = itemType.getSubFieldType(partitioningKeys.get(i)); - typeTraits[i] = TypeTraitProvider.INSTANCE.getTypeTrait(keyType); - } - } - typeTraits[numKeys] = TypeTraitProvider.INSTANCE.getTypeTrait(itemType); - return typeTraits; } public static List<String> getFilterField(Dataset dataset) { diff --git a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java index 73964a9..6d07fc7 100644 --- a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java +++ b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java @@ -29,18 +29,11 @@ import org.apache.asterix.metadata.entities.Dataset; import org.apache.asterix.metadata.entities.Index; import org.apache.asterix.om.types.ARecordType; -import org.apache.asterix.runtime.utils.RuntimeUtils; -import org.apache.hyracks.algebricks.common.constraints.AlgebricksPartitionConstraint; -import org.apache.hyracks.algebricks.common.constraints.AlgebricksPartitionConstraintHelper; import org.apache.hyracks.algebricks.common.exceptions.AlgebricksException; import org.apache.hyracks.algebricks.common.utils.Pair; import org.apache.hyracks.algebricks.core.rewriter.base.PhysicalOptimizationConfig; import org.apache.hyracks.api.dataflow.value.ITypeTraits; import org.apache.hyracks.api.job.JobSpecification; -import org.apache.hyracks.dataflow.std.file.IFileSplitProvider; -import org.apache.hyracks.storage.am.common.dataflow.IIndexDataflowHelperFactory; -import org.apache.hyracks.storage.am.common.dataflow.IndexDropOperatorDescriptor; -import org.apache.hyracks.storage.am.common.dataflow.IndexDataflowHelperFactory; public class IndexUtil { @@ -141,20 +134,6 @@ physicalOptimizationConfig, recType, metaType, enforcedType, enforcedMetaType); secondaryIndexHelper.setExternalFiles(files); return secondaryIndexHelper.buildLoadingJobSpec(); - } - - public static JobSpecification buildDropSecondaryIndexJobSpec(Index index, MetadataProvider metadataProvider, - Dataset dataset) throws AlgebricksException { - JobSpecification spec = RuntimeUtils.createJobSpecification(metadataProvider.getApplicationContext()); - Pair<IFileSplitProvider, AlgebricksPartitionConstraint> splitsAndConstraint = - metadataProvider.getSplitProviderAndConstraints(dataset, index.getIndexName()); - IIndexDataflowHelperFactory indexHelperFactory = new IndexDataflowHelperFactory( - metadataProvider.getStorageComponentProvider().getStorageManager(), splitsAndConstraint.first); - IndexDropOperatorDescriptor btreeDrop = new IndexDropOperatorDescriptor(spec, indexHelperFactory); - AlgebricksPartitionConstraintHelper.setPartitionConstraintInJobSpec(spec, btreeDrop, - splitsAndConstraint.second); - spec.addRoot(btreeDrop); - return spec; } public static JobSpecification buildSecondaryIndexCompactJobSpec(Dataset dataset, Index index, ARecordType recType, -- To view, visit https://asterix-gerrit.ics.uci.edu/1741 To unsubscribe, visit https://asterix-gerrit.ics.uci.edu/settings Gerrit-MessageType: merged Gerrit-Change-Id: I60352d052f684a65fd823a1c7aa33b794bd7f476 Gerrit-PatchSet: 3 Gerrit-Project: asterixdb Gerrit-Branch: master Gerrit-Owner: Yingyi Bu <[email protected]> Gerrit-Reviewer: Jenkins <[email protected]> Gerrit-Reviewer: Till Westmann <[email protected]> Gerrit-Reviewer: Yingyi Bu <[email protected]>
