[24/49] hbase-site git commit: Published site at 3810ba2c6edfc531181ffc9e6c68396a0c2d2027.

2018-09-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/705d69c4/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html
index 025b084..b9f473a 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotManifest.html
@@ -89,504 +89,517 @@
 081  private final ForeignExceptionSnare 
monitor;
 082  private final Configuration conf;
 083  private final Path workingDir;
-084  private final FileSystem fs;
-085  private int manifestSizeLimit;
-086
-087  private SnapshotManifest(final 
Configuration conf, final FileSystem fs,
-088  final Path workingDir, final 
SnapshotDescription desc,
-089  final ForeignExceptionSnare 
monitor) {
-090this.monitor = monitor;
-091this.desc = desc;
-092this.workingDir = workingDir;
-093this.conf = conf;
-094this.fs = fs;
-095
-096this.manifestSizeLimit = 
conf.getInt(SNAPSHOT_MANIFEST_SIZE_LIMIT_CONF_KEY, 64 * 1024 * 1024);
-097  }
-098
-099  /**
-100   * Return a SnapshotManifest instance, 
used for writing a snapshot.
-101   *
-102   * There are two usage pattern:
-103   *  - The Master will create a 
manifest, add the descriptor, offline regions
-104   *and consolidate the snapshot by 
writing all the pending stuff on-disk.
-105   *  manifest = 
SnapshotManifest.create(...)
-106   *  manifest.addRegion(tableDir, 
hri)
-107   *  manifest.consolidate()
-108   *  - The RegionServer will create a 
single region manifest
-109   *  manifest = 
SnapshotManifest.create(...)
-110   *  manifest.addRegion(region)
-111   */
-112  public static SnapshotManifest 
create(final Configuration conf, final FileSystem fs,
-113  final Path workingDir, final 
SnapshotDescription desc,
-114  final ForeignExceptionSnare 
monitor) {
-115return new SnapshotManifest(conf, fs, 
workingDir, desc, monitor);
-116
-117  }
-118
-119  /**
-120   * Return a SnapshotManifest instance 
with the information already loaded in-memory.
-121   *SnapshotManifest manifest = 
SnapshotManifest.open(...)
-122   *TableDescriptor htd = 
manifest.getTableDescriptor()
-123   *for (SnapshotRegionManifest 
regionManifest: manifest.getRegionManifests())
-124   *  hri = 
regionManifest.getRegionInfo()
-125   *  for 
(regionManifest.getFamilyFiles())
-126   *...
-127   */
-128  public static SnapshotManifest 
open(final Configuration conf, final FileSystem fs,
-129  final Path workingDir, final 
SnapshotDescription desc) throws IOException {
-130SnapshotManifest manifest = new 
SnapshotManifest(conf, fs, workingDir, desc, null);
-131manifest.load();
-132return manifest;
-133  }
-134
-135
-136  /**
-137   * Add the table descriptor to the 
snapshot manifest
+084  private final FileSystem rootFs;
+085  private final FileSystem 
workingDirFs;
+086  private int manifestSizeLimit;
+087
+088  /**
+089   *
+090   * @param conf configuration file for 
HBase setup
+091   * @param rootFs root filesystem 
containing HFiles
+092   * @param workingDir file path of where 
the manifest should be located
+093   * @param desc description of snapshot 
being taken
+094   * @param monitor monitor of foreign 
exceptions
+095   * @throws IOException if the working 
directory file system cannot be
+096   * determined from 
the config file
+097   */
+098  private SnapshotManifest(final 
Configuration conf, final FileSystem rootFs,
+099  final Path workingDir, final 
SnapshotDescription desc,
+100  final ForeignExceptionSnare 
monitor) throws IOException {
+101this.monitor = monitor;
+102this.desc = desc;
+103this.workingDir = workingDir;
+104this.conf = conf;
+105this.rootFs = rootFs;
+106this.workingDirFs = 
this.workingDir.getFileSystem(this.conf);
+107this.manifestSizeLimit = 
conf.getInt(SNAPSHOT_MANIFEST_SIZE_LIMIT_CONF_KEY, 64 * 1024 * 1024);
+108  }
+109
+110  /**
+111   * Return a SnapshotManifest instance, 
used for writing a snapshot.
+112   *
+113   * There are two usage pattern:
+114   *  - The Master will create a 
manifest, add the descriptor, offline regions
+115   *and consolidate the snapshot by 
writing all the pending stuff on-disk.
+116   *  manifest = 
SnapshotManifest.create(...)
+117   *  manifest.addRegion(tableDir, 
hri)
+118   *  manifest.consolidate()
+119   *  - The RegionServer will create a 
single region manifest
+120   *  manifest = 
SnapshotManifest.create(...)
+121   *  manifest.addRegion(region)
+122   */
+123  public static SnapshotManifest 
create(final Configuration conf, final FileSystem fs,
+124  final Path workingDir, final 
SnapshotDescription desc,
+125  final 

[24/49] hbase-site git commit: Published site at .

2017-09-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e6bd0774/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
index 2a29b1c..3b0b6dd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
@@ -60,9 +60,9 @@
 052import 
org.apache.hadoop.hbase.procedure2.LockedResource;
 053import 
org.apache.hadoop.hbase.procedure2.Procedure;
 054import 
org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
-055import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-056import 
org.apache.hadoop.hbase.security.User;
-057import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas;
+055import 
org.apache.hadoop.hbase.quotas.GlobalQuotaSettings;
+056import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
+057import 
org.apache.hadoop.hbase.security.User;
 058import 
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;
 059import 
org.apache.yetus.audience.InterfaceAudience;
 060
@@ -1307,598 +1307,608 @@
 1299});
 1300  }
 1301
-1302  public void preSetUserQuota(final 
String user, final Quotas quotas) throws IOException {
-1303execOperation(coprocessors.isEmpty() 
? null : new CoprocessorOperation() {
-1304  @Override
-1305  public void call(MasterObserver 
oserver, ObserverContextMasterCoprocessorEnvironment ctx)
-1306  throws IOException {
-1307oserver.preSetUserQuota(ctx, 
user, quotas);
-1308  }
-1309});
-1310  }
-1311
-1312  public void postSetUserQuota(final 
String user, final Quotas quotas) throws IOException {
-1313execOperation(coprocessors.isEmpty() 
? null : new CoprocessorOperation() {
-1314  @Override
-1315  public void call(MasterObserver 
oserver, ObserverContextMasterCoprocessorEnvironment ctx)
-1316  throws IOException {
-1317oserver.postSetUserQuota(ctx, 
user, quotas);
-1318  }
-1319});
-1320  }
-1321
-1322  public void preSetUserQuota(final 
String user, final TableName table, final Quotas quotas)
-1323  throws IOException {
-1324execOperation(coprocessors.isEmpty() 
? null : new CoprocessorOperation() {
-1325  @Override
-1326  public void call(MasterObserver 
oserver, ObserverContextMasterCoprocessorEnvironment ctx)
-1327  throws IOException {
-1328oserver.preSetUserQuota(ctx, 
user, table, quotas);
-1329  }
-1330});
-1331  }
-1332
-1333  public void postSetUserQuota(final 
String user, final TableName table, final Quotas quotas)
-1334  throws IOException {
-1335execOperation(coprocessors.isEmpty() 
? null : new CoprocessorOperation() {
-1336  @Override
-1337  public void call(MasterObserver 
oserver, ObserverContextMasterCoprocessorEnvironment ctx)
+1302  public void preSetUserQuota(
+1303  final String user, final 
GlobalQuotaSettings quotas) throws IOException {
+1304execOperation(coprocessors.isEmpty() 
? null : new CoprocessorOperation() {
+1305  @Override
+1306  public void call(MasterObserver 
oserver, ObserverContextMasterCoprocessorEnvironment ctx)
+1307  throws IOException {
+1308oserver.preSetUserQuota(ctx, 
user, quotas);
+1309  }
+1310});
+1311  }
+1312
+1313  public void postSetUserQuota(
+1314  final String user, final 
GlobalQuotaSettings quotas) throws IOException {
+1315execOperation(coprocessors.isEmpty() 
? null : new CoprocessorOperation() {
+1316  @Override
+1317  public void call(MasterObserver 
oserver, ObserverContextMasterCoprocessorEnvironment ctx)
+1318  throws IOException {
+1319oserver.postSetUserQuota(ctx, 
user, quotas);
+1320  }
+1321});
+1322  }
+1323
+1324  public void preSetUserQuota(
+1325  final String user, final TableName 
table, final GlobalQuotaSettings quotas)
+1326  throws IOException {
+1327execOperation(coprocessors.isEmpty() 
? null : new CoprocessorOperation() {
+1328  @Override
+1329  public void call(MasterObserver 
oserver, ObserverContextMasterCoprocessorEnvironment ctx)
+1330  throws IOException {
+1331oserver.preSetUserQuota(ctx, 
user, table, quotas);
+1332  }
+1333});
+1334  }
+1335
+1336  public void postSetUserQuota(
+1337  final String user, final TableName 
table, final GlobalQuotaSettings quotas)
 1338  throws IOException {
-1339oserver.postSetUserQuota(ctx, 
user, table, quotas);
-1340  }
-1341});
-1342  }
-1343
-1344  public void preSetUserQuota(final 
String user, final String namespace, final Quotas quotas)

[24/49] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b17bf22b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
index 99a09f9..eb9099e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
@@ -42,2811 +42,2861 @@
 034import java.util.concurrent.TimeUnit;
 035import 
java.util.concurrent.atomic.AtomicReference;
 036import java.util.function.BiConsumer;
-037import java.util.regex.Pattern;
-038import java.util.stream.Collectors;
-039
-040import 
com.google.common.annotations.VisibleForTesting;
-041
-042import io.netty.util.Timeout;
-043import io.netty.util.TimerTask;
+037import java.util.function.Function;
+038import java.util.regex.Pattern;
+039import java.util.stream.Collectors;
+040
+041import 
com.google.common.annotations.VisibleForTesting;
+042import com.google.protobuf.Message;
+043import com.google.protobuf.RpcChannel;
 044
-045import java.util.stream.Stream;
-046
-047import org.apache.commons.io.IOUtils;
-048import org.apache.commons.logging.Log;
-049import 
org.apache.commons.logging.LogFactory;
-050import 
org.apache.hadoop.hbase.ClusterStatus;
-051import 
org.apache.hadoop.hbase.HRegionInfo;
-052import 
org.apache.hadoop.hbase.HRegionLocation;
-053import 
org.apache.hadoop.hbase.MetaTableAccessor;
-054import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-055import 
org.apache.hadoop.hbase.NotServingRegionException;
-056import 
org.apache.hadoop.hbase.ProcedureInfo;
-057import 
org.apache.hadoop.hbase.RegionLoad;
-058import 
org.apache.hadoop.hbase.RegionLocations;
-059import 
org.apache.hadoop.hbase.ServerName;
-060import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-061import 
org.apache.hadoop.hbase.HConstants;
-062import 
org.apache.hadoop.hbase.TableExistsException;
-063import 
org.apache.hadoop.hbase.TableName;
-064import 
org.apache.hadoop.hbase.AsyncMetaTableAccessor;
-065import 
org.apache.hadoop.hbase.TableNotDisabledException;
-066import 
org.apache.hadoop.hbase.TableNotEnabledException;
-067import 
org.apache.hadoop.hbase.TableNotFoundException;
-068import 
org.apache.hadoop.hbase.UnknownRegionException;
-069import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-070import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
-071import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
-072import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-073import 
org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper;
-074import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-075import 
org.apache.hadoop.hbase.client.security.SecurityCapability;
-076import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-077import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-078import 
org.apache.hadoop.hbase.procedure2.LockInfo;
-079import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-080import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-081import 
org.apache.hadoop.hbase.quotas.QuotaTableUtil;
-082import 
org.apache.hadoop.hbase.replication.ReplicationException;
-083import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-084import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-085import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
-086import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-087import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-088import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-089import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
-090import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse;
-091import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
-092import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
-093import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
-094import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
-095import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
-096import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
-098import 

[24/49] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cd7ae54c/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
index 959e3ef..ac32645 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
@@ -1401,6821 +1401,6830 @@
 1393  @Override
 1394  public boolean isSplittable() {
 1395boolean result = isAvailable() 
 !hasReferences();
-1396LOG.info("ASKED IF SPLITTABLE " + 
result, new Throwable("LOGGING"));
-1397return result;
-1398  }
-1399
-1400  @Override
-1401  public boolean isMergeable() {
-1402if (!isAvailable()) {
-1403  LOG.debug("Region " + this
-1404  + " is not mergeable because 
it is closing or closed");
-1405  return false;
-1406}
-1407if (hasReferences()) {
-1408  LOG.debug("Region " + this
-1409  + " is not mergeable because 
it has references");
-1410  return false;
-1411}
-1412
-1413return true;
-1414  }
-1415
-1416  public boolean areWritesEnabled() {
-1417synchronized(this.writestate) {
-1418  return 
this.writestate.writesEnabled;
-1419}
-1420  }
+1396LOG.info("ASKED IF SPLITTABLE " + 
result + " " + getRegionInfo().getShortNameToLog(),
+1397  new Throwable("LOGGING: 
REMOVE"));
+1398// REMOVE BELOW
+1399LOG.info("DEBUG LIST ALL FILES");
+1400for (Store store: 
this.stores.values()) {
+1401  LOG.info("store " + 
store.getColumnFamilyName());
+1402  for (StoreFile sf: 
store.getStorefiles()) {
+1403
LOG.info(sf.toStringDetailed());
+1404  }
+1405}
+1406return result;
+1407  }
+1408
+1409  @Override
+1410  public boolean isMergeable() {
+1411if (!isAvailable()) {
+1412  LOG.debug("Region " + this
+1413  + " is not mergeable because 
it is closing or closed");
+1414  return false;
+1415}
+1416if (hasReferences()) {
+1417  LOG.debug("Region " + this
+1418  + " is not mergeable because 
it has references");
+1419  return false;
+1420}
 1421
-1422  @VisibleForTesting
-1423  public MultiVersionConcurrencyControl 
getMVCC() {
-1424return mvcc;
-1425  }
-1426
-1427  @Override
-1428  public long getMaxFlushedSeqId() {
-1429return maxFlushedSeqId;
-1430  }
-1431
-1432  @Override
-1433  public long 
getReadPoint(IsolationLevel isolationLevel) {
-1434if (isolationLevel != null 
 isolationLevel == IsolationLevel.READ_UNCOMMITTED) {
-1435  // This scan can read even 
uncommitted transactions
-1436  return Long.MAX_VALUE;
-1437}
-1438return mvcc.getReadPoint();
+1422return true;
+1423  }
+1424
+1425  public boolean areWritesEnabled() {
+1426synchronized(this.writestate) {
+1427  return 
this.writestate.writesEnabled;
+1428}
+1429  }
+1430
+1431  @VisibleForTesting
+1432  public MultiVersionConcurrencyControl 
getMVCC() {
+1433return mvcc;
+1434  }
+1435
+1436  @Override
+1437  public long getMaxFlushedSeqId() {
+1438return maxFlushedSeqId;
 1439  }
 1440
 1441  @Override
-1442  public long 
getReadpoint(IsolationLevel isolationLevel) {
-1443return 
getReadPoint(isolationLevel);
-1444  }
-1445
-1446  @Override
-1447  public boolean 
isLoadingCfsOnDemandDefault() {
-1448return 
this.isLoadingCfsOnDemandDefault;
-1449  }
-1450
-1451  /**
-1452   * Close down this HRegion.  Flush the 
cache, shut down each HStore, don't
-1453   * service any more calls.
-1454   *
-1455   * pThis method could take 
some time to execute, so don't call it from a
-1456   * time-sensitive thread.
-1457   *
-1458   * @return Vector of all the storage 
files that the HRegion's component
-1459   * HStores make use of.  It's a list 
of all StoreFile objects. Returns empty
-1460   * vector if already closed and null 
if judged that it should not close.
-1461   *
-1462   * @throws IOException e
-1463   * @throws DroppedSnapshotException 
Thrown when replay of wal is required
-1464   * because a Snapshot was not properly 
persisted. The region is put in closing mode, and the
-1465   * caller MUST abort after this.
-1466   */
-1467  public Mapbyte[], 
ListStoreFile close() throws IOException {
-1468return close(false);
-1469  }
-1470
-1471  private final Object closeLock = new 
Object();
-1472
-1473  /** Conf key for the periodic flush 
interval */
-1474  public static final String 
MEMSTORE_PERIODIC_FLUSH_INTERVAL =
-1475  
"hbase.regionserver.optionalcacheflushinterval";
-1476  /** Default interval for the memstore 
flush */
-1477  public static final int 
DEFAULT_CACHE_FLUSH_INTERVAL = 360;
-1478  /** Default interval for System tables 
memstore flush */
-1479  public 

[24/49] hbase-site git commit: Published site at 4b3e38705cb24aee82615b1b9af47ed549ea1358.

2016-03-03 Thread misty
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4ce8323f/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html
index f21d36e..3761b57 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.Intf.html
@@ -33,10 +33,10 @@
 025  requiredArguments = {
 026@org.jamon.annotations.Argument(name 
= "regionServer", type = "HRegionServer")},
 027  optionalArguments = {
-028@org.jamon.annotations.Argument(name 
= "bcv", type = "String"),
-029@org.jamon.annotations.Argument(name 
= "bcn", type = "String"),
-030@org.jamon.annotations.Argument(name 
= "filter", type = "String"),
-031@org.jamon.annotations.Argument(name 
= "format", type = "String")})
+028@org.jamon.annotations.Argument(name 
= "format", type = "String"),
+029@org.jamon.annotations.Argument(name 
= "bcv", type = "String"),
+030@org.jamon.annotations.Argument(name 
= "bcn", type = "String"),
+031@org.jamon.annotations.Argument(name 
= "filter", type = "String")})
 032public class RSStatusTmpl
 033  extends 
org.jamon.AbstractTemplateProxy
 034{
@@ -77,74 +77,74 @@
 069  return m_regionServer;
 070}
 071private HRegionServer 
m_regionServer;
-072// 24, 1
-073public void setBcv(String bcv)
+072// 22, 1
+073public void setFormat(String 
format)
 074{
-075  // 24, 1
-076  m_bcv = bcv;
-077  m_bcv__IsNotDefault = true;
+075  // 22, 1
+076  m_format = format;
+077  m_format__IsNotDefault = true;
 078}
-079public String getBcv()
+079public String getFormat()
 080{
-081  return m_bcv;
+081  return m_format;
 082}
-083private String m_bcv;
-084public boolean 
getBcv__IsNotDefault()
+083private String m_format;
+084public boolean 
getFormat__IsNotDefault()
 085{
-086  return m_bcv__IsNotDefault;
+086  return m_format__IsNotDefault;
 087}
-088private boolean 
m_bcv__IsNotDefault;
-089// 23, 1
-090public void setBcn(String bcn)
+088private boolean 
m_format__IsNotDefault;
+089// 24, 1
+090public void setBcv(String bcv)
 091{
-092  // 23, 1
-093  m_bcn = bcn;
-094  m_bcn__IsNotDefault = true;
+092  // 24, 1
+093  m_bcv = bcv;
+094  m_bcv__IsNotDefault = true;
 095}
-096public String getBcn()
+096public String getBcv()
 097{
-098  return m_bcn;
+098  return m_bcv;
 099}
-100private String m_bcn;
-101public boolean 
getBcn__IsNotDefault()
+100private String m_bcv;
+101public boolean 
getBcv__IsNotDefault()
 102{
-103  return m_bcn__IsNotDefault;
+103  return m_bcv__IsNotDefault;
 104}
-105private boolean 
m_bcn__IsNotDefault;
-106// 21, 1
-107public void setFilter(String 
filter)
+105private boolean 
m_bcv__IsNotDefault;
+106// 23, 1
+107public void setBcn(String bcn)
 108{
-109  // 21, 1
-110  m_filter = filter;
-111  m_filter__IsNotDefault = true;
+109  // 23, 1
+110  m_bcn = bcn;
+111  m_bcn__IsNotDefault = true;
 112}
-113public String getFilter()
+113public String getBcn()
 114{
-115  return m_filter;
+115  return m_bcn;
 116}
-117private String m_filter;
-118public boolean 
getFilter__IsNotDefault()
+117private String m_bcn;
+118public boolean 
getBcn__IsNotDefault()
 119{
-120  return m_filter__IsNotDefault;
+120  return m_bcn__IsNotDefault;
 121}
-122private boolean 
m_filter__IsNotDefault;
-123// 22, 1
-124public void setFormat(String 
format)
+122private boolean 
m_bcn__IsNotDefault;
+123// 21, 1
+124public void setFilter(String 
filter)
 125{
-126  // 22, 1
-127  m_format = format;
-128  m_format__IsNotDefault = true;
+126  // 21, 1
+127  m_filter = filter;
+128  m_filter__IsNotDefault = true;
 129}
-130public String getFormat()
+130public String getFilter()
 131{
-132  return m_format;
+132  return m_filter;
 133}
-134private String m_format;
-135public boolean 
getFormat__IsNotDefault()
+134private String m_filter;
+135public boolean 
getFilter__IsNotDefault()
 136{
-137  return m_format__IsNotDefault;
+137  return m_filter__IsNotDefault;
 138}
-139private boolean 
m_format__IsNotDefault;
+139private boolean 
m_filter__IsNotDefault;
 140  }
 141  @Override
 142  protected 
org.jamon.AbstractTemplateProxy.ImplData makeImplData()
@@ -156,31 +156,31 @@
 148return (ImplData) 
super.getImplData();
 149  }
 150  
-151  protected String bcv;
-152  public