[24/49] hbase-site git commit: Published site at 3810ba2c6edfc531181ffc9e6c68396a0c2d2027.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/705d69c4/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html index 025b084..b9f473a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html @@ -89,504 +89,517 @@ 081 private final ForeignExceptionSnare monitor; 082 private final Configuration conf; 083 private final Path workingDir; -084 private final FileSystem fs; -085 private int manifestSizeLimit; -086 -087 private SnapshotManifest(final Configuration conf, final FileSystem fs, -088 final Path workingDir, final SnapshotDescription desc, -089 final ForeignExceptionSnare monitor) { -090this.monitor = monitor; -091this.desc = desc; -092this.workingDir = workingDir; -093this.conf = conf; -094this.fs = fs; -095 -096this.manifestSizeLimit = conf.getInt(SNAPSHOT_MANIFEST_SIZE_LIMIT_CONF_KEY, 64 * 1024 * 1024); -097 } -098 -099 /** -100 * Return a SnapshotManifest instance, used for writing a snapshot. -101 * -102 * There are two usage pattern: -103 * - The Master will create a manifest, add the descriptor, offline regions -104 *and consolidate the snapshot by writing all the pending stuff on-disk. -105 * manifest = SnapshotManifest.create(...) -106 * manifest.addRegion(tableDir, hri) -107 * manifest.consolidate() -108 * - The RegionServer will create a single region manifest -109 * manifest = SnapshotManifest.create(...) -110 * manifest.addRegion(region) -111 */ -112 public static SnapshotManifest create(final Configuration conf, final FileSystem fs, -113 final Path workingDir, final SnapshotDescription desc, -114 final ForeignExceptionSnare monitor) { -115return new SnapshotManifest(conf, fs, workingDir, desc, monitor); -116 -117 } -118 -119 /** -120 * Return a SnapshotManifest instance with the information already loaded in-memory. -121 *SnapshotManifest manifest = SnapshotManifest.open(...) -122 *TableDescriptor htd = manifest.getTableDescriptor() -123 *for (SnapshotRegionManifest regionManifest: manifest.getRegionManifests()) -124 * hri = regionManifest.getRegionInfo() -125 * for (regionManifest.getFamilyFiles()) -126 *... -127 */ -128 public static SnapshotManifest open(final Configuration conf, final FileSystem fs, -129 final Path workingDir, final SnapshotDescription desc) throws IOException { -130SnapshotManifest manifest = new SnapshotManifest(conf, fs, workingDir, desc, null); -131manifest.load(); -132return manifest; -133 } -134 -135 -136 /** -137 * Add the table descriptor to the snapshot manifest +084 private final FileSystem rootFs; +085 private final FileSystem workingDirFs; +086 private int manifestSizeLimit; +087 +088 /** +089 * +090 * @param conf configuration file for HBase setup +091 * @param rootFs root filesystem containing HFiles +092 * @param workingDir file path of where the manifest should be located +093 * @param desc description of snapshot being taken +094 * @param monitor monitor of foreign exceptions +095 * @throws IOException if the working directory file system cannot be +096 * determined from the config file +097 */ +098 private SnapshotManifest(final Configuration conf, final FileSystem rootFs, +099 final Path workingDir, final SnapshotDescription desc, +100 final ForeignExceptionSnare monitor) throws IOException { +101this.monitor = monitor; +102this.desc = desc; +103this.workingDir = workingDir; +104this.conf = conf; +105this.rootFs = rootFs; +106this.workingDirFs = this.workingDir.getFileSystem(this.conf); +107this.manifestSizeLimit = conf.getInt(SNAPSHOT_MANIFEST_SIZE_LIMIT_CONF_KEY, 64 * 1024 * 1024); +108 } +109 +110 /** +111 * Return a SnapshotManifest instance, used for writing a snapshot. +112 * +113 * There are two usage pattern: +114 * - The Master will create a manifest, add the descriptor, offline regions +115 *and consolidate the snapshot by writing all the pending stuff on-disk. +116 * manifest = SnapshotManifest.create(...) +117 * manifest.addRegion(tableDir, hri) +118 * manifest.consolidate() +119 * - The RegionServer will create a single region manifest +120 * manifest = SnapshotManifest.create(...) +121 * manifest.addRegion(region) +122 */ +123 public static SnapshotManifest create(final Configuration conf, final FileSystem fs, +124 final Path workingDir, final SnapshotDescription desc, +125 final
[24/49] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e6bd0774/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html index 2a29b1c..3b0b6dd 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html @@ -60,9 +60,9 @@ 052import org.apache.hadoop.hbase.procedure2.LockedResource; 053import org.apache.hadoop.hbase.procedure2.Procedure; 054import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; -055import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -056import org.apache.hadoop.hbase.security.User; -057import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas; +055import org.apache.hadoop.hbase.quotas.GlobalQuotaSettings; +056import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; +057import org.apache.hadoop.hbase.security.User; 058import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription; 059import org.apache.yetus.audience.InterfaceAudience; 060 @@ -1307,598 +1307,608 @@ 1299}); 1300 } 1301 -1302 public void preSetUserQuota(final String user, final Quotas quotas) throws IOException { -1303execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { -1304 @Override -1305 public void call(MasterObserver oserver, ObserverContextMasterCoprocessorEnvironment ctx) -1306 throws IOException { -1307oserver.preSetUserQuota(ctx, user, quotas); -1308 } -1309}); -1310 } -1311 -1312 public void postSetUserQuota(final String user, final Quotas quotas) throws IOException { -1313execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { -1314 @Override -1315 public void call(MasterObserver oserver, ObserverContextMasterCoprocessorEnvironment ctx) -1316 throws IOException { -1317oserver.postSetUserQuota(ctx, user, quotas); -1318 } -1319}); -1320 } -1321 -1322 public void preSetUserQuota(final String user, final TableName table, final Quotas quotas) -1323 throws IOException { -1324execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { -1325 @Override -1326 public void call(MasterObserver oserver, ObserverContextMasterCoprocessorEnvironment ctx) -1327 throws IOException { -1328oserver.preSetUserQuota(ctx, user, table, quotas); -1329 } -1330}); -1331 } -1332 -1333 public void postSetUserQuota(final String user, final TableName table, final Quotas quotas) -1334 throws IOException { -1335execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { -1336 @Override -1337 public void call(MasterObserver oserver, ObserverContextMasterCoprocessorEnvironment ctx) +1302 public void preSetUserQuota( +1303 final String user, final GlobalQuotaSettings quotas) throws IOException { +1304execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { +1305 @Override +1306 public void call(MasterObserver oserver, ObserverContextMasterCoprocessorEnvironment ctx) +1307 throws IOException { +1308oserver.preSetUserQuota(ctx, user, quotas); +1309 } +1310}); +1311 } +1312 +1313 public void postSetUserQuota( +1314 final String user, final GlobalQuotaSettings quotas) throws IOException { +1315execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { +1316 @Override +1317 public void call(MasterObserver oserver, ObserverContextMasterCoprocessorEnvironment ctx) +1318 throws IOException { +1319oserver.postSetUserQuota(ctx, user, quotas); +1320 } +1321}); +1322 } +1323 +1324 public void preSetUserQuota( +1325 final String user, final TableName table, final GlobalQuotaSettings quotas) +1326 throws IOException { +1327execOperation(coprocessors.isEmpty() ? null : new CoprocessorOperation() { +1328 @Override +1329 public void call(MasterObserver oserver, ObserverContextMasterCoprocessorEnvironment ctx) +1330 throws IOException { +1331oserver.preSetUserQuota(ctx, user, table, quotas); +1332 } +1333}); +1334 } +1335 +1336 public void postSetUserQuota( +1337 final String user, final TableName table, final GlobalQuotaSettings quotas) 1338 throws IOException { -1339oserver.postSetUserQuota(ctx, user, table, quotas); -1340 } -1341}); -1342 } -1343 -1344 public void preSetUserQuota(final String user, final String namespace, final Quotas quotas)
[24/49] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b17bf22b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html index 99a09f9..eb9099e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html @@ -42,2811 +42,2861 @@ 034import java.util.concurrent.TimeUnit; 035import java.util.concurrent.atomic.AtomicReference; 036import java.util.function.BiConsumer; -037import java.util.regex.Pattern; -038import java.util.stream.Collectors; -039 -040import com.google.common.annotations.VisibleForTesting; -041 -042import io.netty.util.Timeout; -043import io.netty.util.TimerTask; +037import java.util.function.Function; +038import java.util.regex.Pattern; +039import java.util.stream.Collectors; +040 +041import com.google.common.annotations.VisibleForTesting; +042import com.google.protobuf.Message; +043import com.google.protobuf.RpcChannel; 044 -045import java.util.stream.Stream; -046 -047import org.apache.commons.io.IOUtils; -048import org.apache.commons.logging.Log; -049import org.apache.commons.logging.LogFactory; -050import org.apache.hadoop.hbase.ClusterStatus; -051import org.apache.hadoop.hbase.HRegionInfo; -052import org.apache.hadoop.hbase.HRegionLocation; -053import org.apache.hadoop.hbase.MetaTableAccessor; -054import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -055import org.apache.hadoop.hbase.NotServingRegionException; -056import org.apache.hadoop.hbase.ProcedureInfo; -057import org.apache.hadoop.hbase.RegionLoad; -058import org.apache.hadoop.hbase.RegionLocations; -059import org.apache.hadoop.hbase.ServerName; -060import org.apache.hadoop.hbase.NamespaceDescriptor; -061import org.apache.hadoop.hbase.HConstants; -062import org.apache.hadoop.hbase.TableExistsException; -063import org.apache.hadoop.hbase.TableName; -064import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -065import org.apache.hadoop.hbase.TableNotDisabledException; -066import org.apache.hadoop.hbase.TableNotEnabledException; -067import org.apache.hadoop.hbase.TableNotFoundException; -068import org.apache.hadoop.hbase.UnknownRegionException; -069import org.apache.hadoop.hbase.classification.InterfaceAudience; -070import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -071import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -072import org.apache.hadoop.hbase.client.Scan.ReadType; -073import org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper; -074import org.apache.hadoop.hbase.client.replication.TableCFs; -075import org.apache.hadoop.hbase.client.security.SecurityCapability; -076import org.apache.hadoop.hbase.exceptions.DeserializationException; -077import org.apache.hadoop.hbase.ipc.HBaseRpcController; -078import org.apache.hadoop.hbase.procedure2.LockInfo; -079import org.apache.hadoop.hbase.quotas.QuotaFilter; -080import org.apache.hadoop.hbase.quotas.QuotaSettings; -081import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -082import org.apache.hadoop.hbase.replication.ReplicationException; -083import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -084import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -085import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback; -086import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -087import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -088import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -089import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest; -090import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse; -091import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; -092import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse; -093import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -094import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest; -098import
[24/49] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cd7ae54c/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html index 959e3ef..ac32645 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html @@ -1401,6821 +1401,6830 @@ 1393 @Override 1394 public boolean isSplittable() { 1395boolean result = isAvailable() !hasReferences(); -1396LOG.info("ASKED IF SPLITTABLE " + result, new Throwable("LOGGING")); -1397return result; -1398 } -1399 -1400 @Override -1401 public boolean isMergeable() { -1402if (!isAvailable()) { -1403 LOG.debug("Region " + this -1404 + " is not mergeable because it is closing or closed"); -1405 return false; -1406} -1407if (hasReferences()) { -1408 LOG.debug("Region " + this -1409 + " is not mergeable because it has references"); -1410 return false; -1411} -1412 -1413return true; -1414 } -1415 -1416 public boolean areWritesEnabled() { -1417synchronized(this.writestate) { -1418 return this.writestate.writesEnabled; -1419} -1420 } +1396LOG.info("ASKED IF SPLITTABLE " + result + " " + getRegionInfo().getShortNameToLog(), +1397 new Throwable("LOGGING: REMOVE")); +1398// REMOVE BELOW +1399LOG.info("DEBUG LIST ALL FILES"); +1400for (Store store: this.stores.values()) { +1401 LOG.info("store " + store.getColumnFamilyName()); +1402 for (StoreFile sf: store.getStorefiles()) { +1403 LOG.info(sf.toStringDetailed()); +1404 } +1405} +1406return result; +1407 } +1408 +1409 @Override +1410 public boolean isMergeable() { +1411if (!isAvailable()) { +1412 LOG.debug("Region " + this +1413 + " is not mergeable because it is closing or closed"); +1414 return false; +1415} +1416if (hasReferences()) { +1417 LOG.debug("Region " + this +1418 + " is not mergeable because it has references"); +1419 return false; +1420} 1421 -1422 @VisibleForTesting -1423 public MultiVersionConcurrencyControl getMVCC() { -1424return mvcc; -1425 } -1426 -1427 @Override -1428 public long getMaxFlushedSeqId() { -1429return maxFlushedSeqId; -1430 } -1431 -1432 @Override -1433 public long getReadPoint(IsolationLevel isolationLevel) { -1434if (isolationLevel != null isolationLevel == IsolationLevel.READ_UNCOMMITTED) { -1435 // This scan can read even uncommitted transactions -1436 return Long.MAX_VALUE; -1437} -1438return mvcc.getReadPoint(); +1422return true; +1423 } +1424 +1425 public boolean areWritesEnabled() { +1426synchronized(this.writestate) { +1427 return this.writestate.writesEnabled; +1428} +1429 } +1430 +1431 @VisibleForTesting +1432 public MultiVersionConcurrencyControl getMVCC() { +1433return mvcc; +1434 } +1435 +1436 @Override +1437 public long getMaxFlushedSeqId() { +1438return maxFlushedSeqId; 1439 } 1440 1441 @Override -1442 public long getReadpoint(IsolationLevel isolationLevel) { -1443return getReadPoint(isolationLevel); -1444 } -1445 -1446 @Override -1447 public boolean isLoadingCfsOnDemandDefault() { -1448return this.isLoadingCfsOnDemandDefault; -1449 } -1450 -1451 /** -1452 * Close down this HRegion. Flush the cache, shut down each HStore, don't -1453 * service any more calls. -1454 * -1455 * pThis method could take some time to execute, so don't call it from a -1456 * time-sensitive thread. -1457 * -1458 * @return Vector of all the storage files that the HRegion's component -1459 * HStores make use of. It's a list of all StoreFile objects. Returns empty -1460 * vector if already closed and null if judged that it should not close. -1461 * -1462 * @throws IOException e -1463 * @throws DroppedSnapshotException Thrown when replay of wal is required -1464 * because a Snapshot was not properly persisted. The region is put in closing mode, and the -1465 * caller MUST abort after this. -1466 */ -1467 public Mapbyte[], ListStoreFile close() throws IOException { -1468return close(false); -1469 } -1470 -1471 private final Object closeLock = new Object(); -1472 -1473 /** Conf key for the periodic flush interval */ -1474 public static final String MEMSTORE_PERIODIC_FLUSH_INTERVAL = -1475 "hbase.regionserver.optionalcacheflushinterval"; -1476 /** Default interval for the memstore flush */ -1477 public static final int DEFAULT_CACHE_FLUSH_INTERVAL = 360; -1478 /** Default interval for System tables memstore flush */ -1479 public
[24/49] hbase-site git commit: Published site at 4b3e38705cb24aee82615b1b9af47ed549ea1358.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4ce8323f/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html index f21d36e..3761b57 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html @@ -33,10 +33,10 @@ 025 requiredArguments = { 026@org.jamon.annotations.Argument(name = "regionServer", type = "HRegionServer")}, 027 optionalArguments = { -028@org.jamon.annotations.Argument(name = "bcv", type = "String"), -029@org.jamon.annotations.Argument(name = "bcn", type = "String"), -030@org.jamon.annotations.Argument(name = "filter", type = "String"), -031@org.jamon.annotations.Argument(name = "format", type = "String")}) +028@org.jamon.annotations.Argument(name = "format", type = "String"), +029@org.jamon.annotations.Argument(name = "bcv", type = "String"), +030@org.jamon.annotations.Argument(name = "bcn", type = "String"), +031@org.jamon.annotations.Argument(name = "filter", type = "String")}) 032public class RSStatusTmpl 033 extends org.jamon.AbstractTemplateProxy 034{ @@ -77,74 +77,74 @@ 069 return m_regionServer; 070} 071private HRegionServer m_regionServer; -072// 24, 1 -073public void setBcv(String bcv) +072// 22, 1 +073public void setFormat(String format) 074{ -075 // 24, 1 -076 m_bcv = bcv; -077 m_bcv__IsNotDefault = true; +075 // 22, 1 +076 m_format = format; +077 m_format__IsNotDefault = true; 078} -079public String getBcv() +079public String getFormat() 080{ -081 return m_bcv; +081 return m_format; 082} -083private String m_bcv; -084public boolean getBcv__IsNotDefault() +083private String m_format; +084public boolean getFormat__IsNotDefault() 085{ -086 return m_bcv__IsNotDefault; +086 return m_format__IsNotDefault; 087} -088private boolean m_bcv__IsNotDefault; -089// 23, 1 -090public void setBcn(String bcn) +088private boolean m_format__IsNotDefault; +089// 24, 1 +090public void setBcv(String bcv) 091{ -092 // 23, 1 -093 m_bcn = bcn; -094 m_bcn__IsNotDefault = true; +092 // 24, 1 +093 m_bcv = bcv; +094 m_bcv__IsNotDefault = true; 095} -096public String getBcn() +096public String getBcv() 097{ -098 return m_bcn; +098 return m_bcv; 099} -100private String m_bcn; -101public boolean getBcn__IsNotDefault() +100private String m_bcv; +101public boolean getBcv__IsNotDefault() 102{ -103 return m_bcn__IsNotDefault; +103 return m_bcv__IsNotDefault; 104} -105private boolean m_bcn__IsNotDefault; -106// 21, 1 -107public void setFilter(String filter) +105private boolean m_bcv__IsNotDefault; +106// 23, 1 +107public void setBcn(String bcn) 108{ -109 // 21, 1 -110 m_filter = filter; -111 m_filter__IsNotDefault = true; +109 // 23, 1 +110 m_bcn = bcn; +111 m_bcn__IsNotDefault = true; 112} -113public String getFilter() +113public String getBcn() 114{ -115 return m_filter; +115 return m_bcn; 116} -117private String m_filter; -118public boolean getFilter__IsNotDefault() +117private String m_bcn; +118public boolean getBcn__IsNotDefault() 119{ -120 return m_filter__IsNotDefault; +120 return m_bcn__IsNotDefault; 121} -122private boolean m_filter__IsNotDefault; -123// 22, 1 -124public void setFormat(String format) +122private boolean m_bcn__IsNotDefault; +123// 21, 1 +124public void setFilter(String filter) 125{ -126 // 22, 1 -127 m_format = format; -128 m_format__IsNotDefault = true; +126 // 21, 1 +127 m_filter = filter; +128 m_filter__IsNotDefault = true; 129} -130public String getFormat() +130public String getFilter() 131{ -132 return m_format; +132 return m_filter; 133} -134private String m_format; -135public boolean getFormat__IsNotDefault() +134private String m_filter; +135public boolean getFilter__IsNotDefault() 136{ -137 return m_format__IsNotDefault; +137 return m_filter__IsNotDefault; 138} -139private boolean m_format__IsNotDefault; +139private boolean m_filter__IsNotDefault; 140 } 141 @Override 142 protected org.jamon.AbstractTemplateProxy.ImplData makeImplData() @@ -156,31 +156,31 @@ 148return (ImplData) super.getImplData(); 149 } 150 -151 protected String bcv; -152 public