Repository: accumulo Updated Branches: refs/heads/asf-site 82bf9441c -> c0655661f refs/heads/gh-pages b873c76e8 -> 358b7b452
http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/Table_Design.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/Table_Design.md b/user_manual_1.3-incubating/Table_Design.md deleted file mode 100644 index f5dc981..0000000 --- a/user_manual_1.3-incubating/Table_Design.md +++ /dev/null @@ -1,197 +0,0 @@ ---- -title: "User Manual: Table Design" ---- - -** Next:** [High-Speed Ingest][2] ** Up:** [Apache Accumulo User Manual Version 1.3][4] ** Previous:** [Table Configuration][6] ** [Contents][8]** - -<a id="CHILD_LINKS"></a>**Subsections** - -* [Basic Table][9] -* [RowID Design][10] -* [Indexing][11] -* [Entity-Attribute and Graph Tables][12] -* [Document-Partitioned Indexing][13] - -* * * - -## <a id="Table_Design"></a> Table Design - -## <a id="Basic_Table"></a> Basic Table - -Since Accumulo tables are sorted by row ID, each table can be thought of as being indexed by the row ID. Lookups performed row ID can be executed quickly, by doing a binary search, first across the tablets, and then within a tablet. Clients should choose a row ID carefully in order to support their desired application. A simple rule is to select a unique identifier as the row ID for each entity to be stored and assign all the other attributes to be tracked to be columns under this row ID. For example, if we have the following data in a comma-separated file: - - - userid,age,address,account-balance - - -We might choose to store this data using the userid as the rowID and the rest of the data in column families: - - - Mutation m = new Mutation(new Text(userid)); - m.put(new Text("age"), age); - m.put(new Text("address"), address); - m.put(new Text("balance"), account_balance); - - writer.add(m); - - -We could then retrieve any of the columns for a specific userid by specifying the userid as the range of a scanner and fetching specific columns: - - - Range r = new Range(userid, userid); // single row - Scanner s = conn.createScanner("userdata", auths); - s.setRange(r); - s.fetchColumnFamily(new Text("age")); - - for(Entry<Key,Value> entry : s) - System.out.println(entry.getValue().toString()); - - -## <a id="RowID_Design"></a> RowID Design - -Often it is necessary to transform the rowID in order to have rows ordered in a way that is optimal for anticipated access patterns. A good example of this is reversing the order of components of internet domain names in order to group rows of the same parent domain together: - - - com.google.code - com.google.labs - com.google.mail - com.yahoo.mail - com.yahoo.research - - -Some data may result in the creation of very large rows - rows with many columns. In this case the table designer may wish to split up these rows for better load balancing while keeping them sorted together for scanning purposes. This can be done by appending a random substring at the end of the row: - - - com.google.code_00 - com.google.code_01 - com.google.code_02 - com.google.labs_00 - com.google.mail_00 - com.google.mail_01 - - -It could also be done by adding a string representation of some period of time such as date to the week or month: - - - com.google.code_201003 - com.google.code_201004 - com.google.code_201005 - com.google.labs_201003 - com.google.mail_201003 - com.google.mail_201004 - - -Appending dates provides the additional capability of restricting a scan to a given date range. - -## <a id="Indexing"></a> Indexing - -In order to support lookups via more than one attribute of an entity, additional indexes can be built. However, because Accumulo tables can support any number of columns without specifying them beforehand, a single additional index will often suffice for supporting lookups of records in the main table. Here, the index has, as the rowID, the Value or Term from the main table, the column families are the same, and the column qualifier of the index table contains the rowID from the main table. - -![converted table][14] - -Note: We store rowIDs in the column qualifier rather than the Value so that we can have more than one rowID associated with a particular term within the index. If we stored this in the Value we would only see one of the rows in which the value appears since Accumulo is configured by default to return the one most recent value associated with a key. - -Lookups can then be done by scanning the Index Table first for occurrences of the desired values in the columns specified, which returns a list of row ID from the main table. These can then be used to retrieve each matching record, in their entirety, or a subset of their columns, from the Main Table. - -To support efficient lookups of multiple rowIDs from the same table, the Accumulo client library provides a BatchScanner. Users specify a set of Ranges to the BatchScanner, which performs the lookups in multiple threads to multiple servers and returns an Iterator over all the rows retrieved. The rows returned are NOT in sorted order, as is the case with the basic Scanner interface. - - - // first we scan the index for IDs of rows matching our query - - Text term = new Text("mySearchTerm"); - - HashSet<Text> matchingRows = new HashSet<Text>(); - - Scanner indexScanner = createScanner("index", auths); - indexScanner.setRange(new Range(term, term)); - - // we retrieve the matching rowIDs and create a set of ranges - for(Entry<Key,Value> entry : indexScanner) - matchingRows.add(new Text(entry.getValue())); - - // now we pass the set of rowIDs to the batch scanner to retrieve them - BatchScanner bscan = conn.createBatchScanner("table", auths, 10); - - bscan.setRanges(matchingRows); - bscan.fetchFamily("attributes"); - - for(Entry<Key,Value> entry : scan) - System.out.println(e.getValue()); - - -One advantage of the dynamic schema capabilities of Accumulo is that different fields may be indexed into the same physical table. However, it may be necessary to create different index tables if the terms must be formatted differently in order to maintain proper sort order. For example, real numbers must be formatted differently than their usual notation in order to be sorted correctly. In these cases, usually one index per unique data type will suffice. - -## <a id="Entity-Attribute_and_Graph_Tables"></a> Entity-Attribute and Graph Tables - -Accumulo is ideal for storing entities and their attributes, especially of the attributes are sparse. It is often useful to join several datasets together on common entities within the same table. This can allow for the representation of graphs, including nodes, their attributes, and connections to other nodes. - -Rather than storing individual events, Entity-Attribute or Graph tables store aggregate information about the entities involved in the events and the relationships between entities. This is often preferrable when single events aren't very useful and when a continuously updated summarization is desired. - -The physical shema for an entity-attribute or graph table is as follows: - -![converted table][15] - -For example, to keep track of employees, managers and products the following entity-attribute table could be used. Note that the weights are not always necessary and are set to 0 when not used. - -![converted table][16] - - - -To allow efficient updating of edge weights, an aggregating iterator can be configured to add the value of all mutations applied with the same key. These types of tables can easily be created from raw events by simply extracting the entities, attributes, and relationships from individual events and inserting the keys into Accumulo each with a count of 1. The aggregating iterator will take care of maintaining the edge weights. - -## <a id="Document-Partitioned_Indexing"></a> Document-Partitioned Indexing - -Using a simple index as described above works well when looking for records that match one of a set of given criteria. When looking for records that match more than one criterion simultaneously, such as when looking for documents that contain all of the words `the' and `white' and `house', there are several issues. - -First is that the set of all records matching any one of the search terms must be sent to the client, which incurs a lot of network traffic. The second problem is that the client is responsible for performing set intersection on the sets of records returned to eliminate all but the records matching all search terms. The memory of the client may easily be overwhelmed during this operation. - -For these reasons Accumulo includes support for a scheme known as sharded indexing, in which these set operations can be performed at the TabletServers and decisions about which records to include in the result set can be made without incurring network traffic. - -This is accomplished via partitioning records into bins that each reside on at most one TabletServer, and then creating an index of terms per record within each bin as follows: - -![converted table][17] - -Documents or records are mapped into bins by a user-defined ingest application. By storing the BinID as the RowID we ensure that all the information for a particular bin is contained in a single tablet and hosted on a single TabletServer since Accumulo never splits rows across tablets. Storing the Terms as column families serves to enable fast lookups of all the documents within this bin that contain the given term. - -Finally, we perform set intersection operations on the TabletServer via a special iterator called the Intersecting Iterator. Since documents are partitioned into many bins, a search of all documents must search every bin. We can use the BatchScanner to scan all bins in parallel. The Intersecting Iterator should be enabled on a BatchScanner within user query code as follows: - - - Text[] terms = {new Text("the"), new Text("white"), new Text("house")}; - - BatchScanner bs = conn.createBatchScanner(table, auths, 20); - bs.setScanIterators(20, IntersectingIterator.class.getName(), "ii"); - - // tells scanner to look for terms in the column family and sends terms - bs.setScanIteratorOption("ii", - IntersectingIterator.columnFamiliesOptionName, - IntersectingIterator.encodeColumns(terms)); - - bs.setRanges(Collections.singleton(new Range())); - - for(Entry<Key,Value> entry : bs) { - System.out.println(" " + entry.getKey().getColumnQualifier()); - } - - -This code effectively has the BatchScanner scan all tablets of a table, looking for documents that match all the given terms. Because all tablets are being scanned for every query, each query is more expensive than other Accumulo scans, which typically involve a small number of TabletServers. This reduces the number of concurrent queries supported and is subject to what is known as the `straggler' problem in which every query runs as slow as the slowest server participating. - -Of course, fast servers will return their results to the client which can display them to the user immediately while they wait for the rest of the results to arrive. If the results are unordered this is quite effective as the first results to arrive are as good as any others to the user. - -* * * - -** Next:** [High-Speed Ingest][2] ** Up:** [Apache Accumulo User Manual Version 1.3][4] ** Previous:** [Table Configuration][6] ** [Contents][8]** - -[2]: High_Speed_Ingest.html -[4]: accumulo_user_manual.html -[6]: Table_Configuration.html -[8]: Contents.html -[9]: Table_Design.html#Basic_Table -[10]: Table_Design.html#RowID_Design -[11]: Table_Design.html#Indexing -[12]: Table_Design.html#Entity-Attribute_and_Graph_Tables -[13]: Table_Design.html#Document-Partitioned_Indexing -[14]: img2.png -[15]: img3.png -[16]: img4.png -[17]: img5.png - http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/Writing_Accumulo_Clients.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/Writing_Accumulo_Clients.md b/user_manual_1.3-incubating/Writing_Accumulo_Clients.md deleted file mode 100644 index a31f91c..0000000 --- a/user_manual_1.3-incubating/Writing_Accumulo_Clients.md +++ /dev/null @@ -1,124 +0,0 @@ ---- -title: "User Manual: Writing Accumulo Clients" ---- - -** Next:** [Table Configuration][2] ** Up:** [Apache Accumulo User Manual Version 1.3][4] ** Previous:** [Accumulo Shell][6] ** [Contents][8]** - -<a id="CHILD_LINKS"></a>**Subsections** - -* [Writing Data][9] -* [Reading Data][10] - -* * * - -## <a id="Writing_Accumulo_Clients"></a> Writing Accumulo Clients - -All clients must first identify the Accumulo instance to which they will be communicating. Code to do this is as follows: - - - String instanceName = "myinstance"; - String zooServers = "zooserver-one,zooserver-two" - Instance inst = new ZooKeeperInstance(instanceName, zooServers); - - Connector conn = new Connector(inst, "user","passwd".getBytes()); - - -## <a id="Writing_Data"></a> Writing Data - -Data are written to Accumulo by creating Mutation objects that represent all the changes to the columns of a single row. The changes are made atomically in the TabletServer. Clients then add Mutations to a BatchWriter which submits them to the appropriate TabletServers. - -Mutations can be created thus: - - - Text rowID = new Text("row1"); - Text colFam = new Text("myColFam"); - Text colQual = new Text("myColQual"); - ColumnVisibility colVis = new ColumnVisibility("public"); - long timestamp = System.currentTimeMillis(); - - Value value = new Value("myValue".getBytes()); - - Mutation mutation = new Mutation(rowID); - mutation.put(colFam, colQual, colVis, timestamp, value); - - -### <a id="BatchWriter"></a> BatchWriter - -The BatchWriter is highly optimized to send Mutations to multiple TabletServers and automatically batches Mutations destined for the same TabletServer to amortize network overhead. Care must be taken to avoid changing the contents of any Object passed to the BatchWriter since it keeps objects in memory while batching. - -Mutations are added to a BatchWriter thus: - - - long memBuf = 1000000L; // bytes to store before sending a batch - long timeout = 1000L; // milliseconds to wait before sending - int numThreads = 10; - - BatchWriter writer = - conn.createBatchWriter("table", memBuf, timeout, numThreads) - - writer.add(mutation); - - writer.close(); - - -An example of using the batch writer can be found at -accumulo/docs/examples/README.batch - -## <a id="Reading_Data"></a> Reading Data - -Accumulo is optimized to quickly retrieve the value associated with a given key, and to efficiently return ranges of consecutive keys and their associated values. - -### <a id="Scanner"></a> Scanner - -To retrieve data, Clients use a Scanner, which provides acts like an Iterator over keys and values. Scanners can be configured to start and stop at particular keys, and to return a subset of the columns available. - - - // specify which visibilities we are allowed to see - Authorizations auths = new Authorizations("public"); - - Scanner scan = - conn.createScanner("table", auths); - - scan.setRange(new Range("harry","john")); - scan.fetchFamily("attributes"); - - for(Entry<Key,Value> entry : scan) { - String row = e.getKey().getRow(); - Value value = e.getValue(); - } - - -### <a id="BatchScanner"></a> BatchScanner - -For some types of access, it is more efficient to retrieve several ranges simultaneously. This arises when accessing a set of rows that are not consecutive whose IDs have been retrieved from a secondary index, for example. - -The BatchScanner is configured similarly to the Scanner; it can be configured to retrieve a subset of the columns available, but rather than passing a single Range, BatchScanners accept a set of Ranges. It is important to note that the keys returned by a BatchScanner are not in sorted order since the keys streamed are from multiple TabletServers in parallel. - - - ArrayList<Range> ranges = new ArrayList<Range>(); - // populate list of ranges ... - - BatchScanner bscan = - conn.createBatchScanner("table", auths, 10); - - bscan.setRanges(ranges); - bscan.fetchFamily("attributes"); - - for(Entry<Key,Value> entry : scan) - System.out.println(e.getValue()); - - -An example of the BatchScanner can be found at -accumulo/docs/examples/README.batch - -* * * - -** Next:** [Table Configuration][2] ** Up:** [Apache Accumulo User Manual Version 1.3][4] ** Previous:** [Accumulo Shell][6] ** [Contents][8]** - -[2]: Table_Configuration.html -[4]: accumulo_user_manual.html -[6]: Accumulo_Shell.html -[8]: Contents.html -[9]: Writing_Accumulo_Clients.html#Writing_Data -[10]: Writing_Accumulo_Clients.html#Reading_Data - http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/accumulo_user_manual.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/accumulo_user_manual.md b/user_manual_1.3-incubating/accumulo_user_manual.md deleted file mode 100644 index e40cda8..0000000 --- a/user_manual_1.3-incubating/accumulo_user_manual.md +++ /dev/null @@ -1,49 +0,0 @@ ---- -title: "User Manual: index" ---- - -** Next:** [Contents][2] ** [Contents][2]** - - - -## Apache Accumulo User Manual -Version 1.3 - - - - -* * * - -<a id="CHILD_LINKS"></a> - -* [Contents][2] -* [Introduction][6] -* [Accumulo Design][7] -* [Accumulo Shell][8] -* [Writing Accumulo Clients][9] -* [Table Configuration][10] -* [Table Design][11] -* [High-Speed Ingest][12] -* [Analytics][13] -* [Security][14] -* [Administration][15] -* [Shell Commands][16] - - - - -* * * - -[2]: Contents.html -[6]: Introduction.html -[7]: Accumulo_Design.html -[8]: Accumulo_Shell.html -[9]: Writing_Accumulo_Clients.html -[10]: Table_Configuration.html -[11]: Table_Design.html -[12]: High_Speed_Ingest.html -[13]: Analytics.html -[14]: Security.html -[15]: Administration.html -[16]: Shell_Commands.html - http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/data_distribution.png ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/data_distribution.png b/user_manual_1.3-incubating/data_distribution.png deleted file mode 100644 index 71b585b..0000000 Binary files a/user_manual_1.3-incubating/data_distribution.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples.md b/user_manual_1.3-incubating/examples.md deleted file mode 100644 index 4654a0a..0000000 --- a/user_manual_1.3-incubating/examples.md +++ /dev/null @@ -1,7 +0,0 @@ ---- -title: Examples -redirect_to: examples/ ---- - -This page redirects to the 1.3 examples - http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/aggregation.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/aggregation.md b/user_manual_1.3-incubating/examples/aggregation.md deleted file mode 100755 index 3def5e4..0000000 --- a/user_manual_1.3-incubating/examples/aggregation.md +++ /dev/null @@ -1,36 +0,0 @@ ---- -title: Aggregation Example ---- - -This is a simple aggregation example. To build this example run maven and then -copy the produced jar into the accumulo lib dir. This is already done in the -tar distribution. - - $ bin/accumulo shell -u username - Enter current password for 'username'@'instance': *** - - Shell - Apache Accumulo Interactive Shell - - - - version: 1.3.x-incubating - - instance name: instance - - instance id: 00000000-0000-0000-0000-000000000000 - - - - type 'help' for a list of available commands - - - username@instance> createtable aggtest1 -a app=org.apache.accumulo.examples.aggregation.SortedSetAggregator - username@instance aggtest1> insert foo app 1 a - username@instance aggtest1> insert foo app 1 b - username@instance aggtest1> scan - foo app:1 [] a,b - username@instance aggtest1> insert foo app 1 z,1,foo,w - username@instance aggtest1> scan - foo app:1 [] 1,a,b,foo,w,z - username@instance aggtest1> insert foo app 2 cat,dog,muskrat - username@instance aggtest1> insert foo app 2 mouse,bird - username@instance aggtest1> scan - foo app:1 [] 1,a,b,foo,w,z - foo app:2 [] bird,cat,dog,mouse,muskrat - username@instance aggtest1> - -In this example a table is created and the example set aggregator is -applied to the column family app. http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/batch.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/batch.md b/user_manual_1.3-incubating/examples/batch.md deleted file mode 100755 index 6989f59..0000000 --- a/user_manual_1.3-incubating/examples/batch.md +++ /dev/null @@ -1,39 +0,0 @@ ---- -title: Batch Writing and Scanning Example ---- - -This is an example of how to use the batch writer and batch scanner. To compile -the example, run maven and copy the produced jar into the accumulo lib dir. -This is already done in the tar distribution. - -Below are commands that add 10000 entries to accumulo and then do 100 random -queries. The write command generates random 50 byte values. - -Be sure to use the name of your instance (given as instance here) and the appropriate -list of zookeeper nodes (given as zookeepers here). - -Before you run this, you must ensure that the user you are running has the -"exampleVis" authorization. (you can set this in the shell with "setauths -u username -s exampleVis") - - $ ./bin/accumulo shell -u root - > setauths -u username -s exampleVis - > exit - -You must also create the table, batchtest1, ahead of time. (In the shell, use "createtable batchtest1") - - $ ./bin/accumulo shell -u username - > createtable batchtest1 - > exit - $ ./bin/accumulo org.apache.accumulo.examples.client.SequentialBatchWriter instance zookeepers username password batchtest1 0 10000 50 20000000 500 20 exampleVis - $ ./bin/accumulo org.apache.accumulo.examples.client.RandomBatchScanner instance zookeepers username password batchtest1 100 0 10000 50 20 exampleVis - 07 11:33:11,103 [client.CountingVerifyingReceiver] INFO : Generating 100 random queries... - 07 11:33:11,112 [client.CountingVerifyingReceiver] INFO : finished - 07 11:33:11,260 [client.CountingVerifyingReceiver] INFO : 694.44 lookups/sec 0.14 secs - - 07 11:33:11,260 [client.CountingVerifyingReceiver] INFO : num results : 100 - - 07 11:33:11,364 [client.CountingVerifyingReceiver] INFO : Generating 100 random queries... - 07 11:33:11,370 [client.CountingVerifyingReceiver] INFO : finished - 07 11:33:11,416 [client.CountingVerifyingReceiver] INFO : 2173.91 lookups/sec 0.05 secs - - 07 11:33:11,416 [client.CountingVerifyingReceiver] INFO : num results : 100 http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/bloom.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/bloom.md b/user_manual_1.3-incubating/examples/bloom.md deleted file mode 100755 index d77a966..0000000 --- a/user_manual_1.3-incubating/examples/bloom.md +++ /dev/null @@ -1,113 +0,0 @@ ---- -title: Bloom Filter Example ---- - -This example shows how to create a table with bloom filters enabled. It also -shows how bloom filters increase query performance when looking for values that -do not exist in a table. - -Below table named bloom_test is created and bloom filters are enabled. - - $ ./accumulo shell -u username -p password - Shell - Apache Accumulo Interactive Shell - - version: 1.3.x-incubating - - instance name: instance - - instance id: 00000000-0000-0000-0000-000000000000 - - - - type 'help' for a list of available commands - - - username@instance> setauths -u username -s exampleVis - username@instance> createtable bloom_test - username@instance bloom_test> config -t bloom_test -s table.bloom.enabled=true - username@instance bloom_test> exit - -Below 1 million random values are inserted into accumulo. The randomly -generated rows range between 0 and 1 billion. The random number generator is -initialized with the seed 7. - - $ ./bin/accumulo org.apache.accumulo.examples.client.RandomBatchWriter -s 7 instance zookeepers username password bloom_test 1000000 0 1000000000 50 2000000 60000 3 exampleVis - -Below the table is flushed, look at the monitor page and wait for the flush to -complete. - - $ ./bin/accumulo shell -u username -p password - username@instance> flush -t bloom_test - Flush of table bloom_test initiated... - username@instance> exit - -The flush will be finished when there are no entries in memory and the -number of minor compactions goes to zero. Refresh the page to see changes to the table. - -After the flush completes, 500 random queries are done against the table. The -same seed is used to generate the queries, therefore everything is found in the -table. - - $ ./bin/accumulo org.apache.accumulo.examples.client.RandomBatchScanner -s 7 instance zookeepers username password bloom_test 500 0 1000000000 50 20 exampleVis - Generating 500 random queries...finished - 96.19 lookups/sec 5.20 secs - num results : 500 - Generating 500 random queries...finished - 102.35 lookups/sec 4.89 secs - num results : 500 - -Below another 500 queries are performed, using a different seed which results -in nothing being found. In this case the lookups are much faster because of -the bloom filters. - - $ ../bin/accumulo org.apache.accumulo.examples.client.RandomBatchScanner -s 8 instance zookeepers username password bloom_test 500 0 1000000000 50 20 exampleVis - Generating 500 random queries...finished - 2212.39 lookups/sec 0.23 secs - num results : 0 - Did not find 500 rows - Generating 500 random queries...finished - 4464.29 lookups/sec 0.11 secs - num results : 0 - Did not find 500 rows - -******************************************************************************** - -Bloom filters can also speed up lookups for entries that exist. In accumulo -data is divided into tablets and each tablet has multiple map files. Every -lookup in accumulo goes to a specific tablet where a lookup is done on each -map file in the tablet. So if a tablet has three map files, lookup performance -can be three times slower than a tablet with one map file. However if the map -files contain unique sets of data, then bloom filters can help eliminate map -files that do not contain the row being looked up. To illustrate this two -identical tables were created using the following process. One table had bloom -filters, the other did not. Also the major compaction ratio was increased to -prevent the files from being compacted into one file. - - * Insert 1 million entries using RandomBatchWriter with a seed of 7 - * Flush the table using the shell - * Insert 1 million entries using RandomBatchWriter with a seed of 8 - * Flush the table using the shell - * Insert 1 million entries using RandomBatchWriter with a seed of 9 - * Flush the table using the shell - -After following the above steps, each table will have a tablet with three map -files. Each map file will contain 1 million entries generated with a different -seed. - -Below 500 lookups are done against the table without bloom filters using random -NG seed 7. Even though only one map file will likely contain entries for this -seed, all map files will be interrogated. - - $ ./bin/accumulo org.apache.accumulo.examples.client.RandomBatchScanner -s 7 instance zookeepers username password bloom_test1 500 0 1000000000 50 20 exampleVis - Generating 500 random queries...finished - 35.09 lookups/sec 14.25 secs - num results : 500 - Generating 500 random queries...finished - 35.33 lookups/sec 14.15 secs - num results : 500 - -Below the same lookups are done against the table with bloom filters. The -lookups were 2.86 times faster because only one map file was used, even though three -map files existed. - - $ ./bin/accumulo org.apache.accumulo.examples.client.RandomBatchScanner -s 7 instance zookeepers username password bloom_test2 500 0 1000000000 50 20 exampleVis - Generating 500 random queries...finished - 99.03 lookups/sec 5.05 secs - num results : 500 - Generating 500 random queries...finished - 101.15 lookups/sec 4.94 secs - num results : 500 http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/bulkIngest.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/bulkIngest.md b/user_manual_1.3-incubating/examples/bulkIngest.md deleted file mode 100755 index 9b3081c..0000000 --- a/user_manual_1.3-incubating/examples/bulkIngest.md +++ /dev/null @@ -1,20 +0,0 @@ ---- -title: Bulk Ingest Example ---- - -This is an example of how to bulk ingest data into accumulo using map reduce. - -The following commands show how to run this example. This example creates a -table called test_bulk which has two initial split points. Then 1000 rows of -test data are created in HDFS. After that the 1000 rows are ingested into -accumulo. Then we verify the 1000 rows are in accumulo. The -first two arguments to all of the commands except for GenerateTestData are the -accumulo instance name, and a comma-separated list of zookeepers. - - $ ./bin/accumulo org.apache.accumulo.examples.mapreduce.bulk.SetupTable instance zookeepers username password test_bulk row_00000333 row_00000666 - $ ./bin/accumulo org.apache.accumulo.examples.mapreduce.bulk.GenerateTestData 0 1000 bulk/test_1.txt - - $ ./bin/tool.sh lib/accumulo-examples-*[^c].jar org.apache.accumulo.examples.mapreduce.bulk.BulkIngestExample instance zookeepers username password test_bulk bulk tmp/bulkWork - $ ./bin/accumulo org.apache.accumulo.examples.mapreduce.bulk.VerifyIngest instance zookeepers username password test_bulk 0 1000 - -For a high level discussion of bulk ingest, see the docs dir. http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/constraints.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/constraints.md b/user_manual_1.3-incubating/examples/constraints.md deleted file mode 100755 index 58aa354..0000000 --- a/user_manual_1.3-incubating/examples/constraints.md +++ /dev/null @@ -1,34 +0,0 @@ ---- -title: Constraints Example ---- - -This an example of how to create a table with constraints. Below a table is -create with two example constraints. One constraints does not allow non alpha -numeric keys. The other constraint does not allow non numeric values. Two -inserts that violate these constraints are attempted and denied. The scan at -the end shows the inserts were not allowed. - - $ ./bin/accumulo shell -u username -p pass - - Shell - Apache Accumulo Interactive Shell - - - - version: 1.3.x-incubating - - instance name: instance - - instance id: 00000000-0000-0000-0000-000000000000 - - - - type 'help' for a list of available commands - - - username@instance> createtable testConstraints - username@instance testConstraints> config -t testConstraints -s table.constraint.1=org.apache.accumulo.examples.constraints.NumericValueConstraint - username@instance testConstraints> config -t testConstraints -s table.constraint.2=org.apache.accumulo.examples.constraints.AlphaNumKeyConstrain - username@instance testConstraints> insert r1 cf1 cq1 1111 - username@instance testConstraints> insert r1 cf1 cq1 ABC - Constraint Failures: - ConstraintViolationSummary(constrainClass:org.apache.accumulo.examples.constraints.NumericValueConstraint, violationCode:1, violationDescription:Value is not numeric, numberOfViolatingMutations:1) - username@instance testConstraints> insert r1! cf1 cq1 ABC - Constraint Failures: - ConstraintViolationSummary(constrainClass:org.apache.accumulo.examples.constraints.NumericValueConstraint, violationCode:1, violationDescription:Value is not numeric, numberOfViolatingMutations:1) - ConstraintViolationSummary(constrainClass:org.apache.accumulo.examples.constraints.AlphaNumKeyConstraint, violationCode:1, violationDescription:Row was not alpha numeric, numberOfViolatingMutations:1) - username@instance testConstraints> scan - r1 cf1:cq1 [] 1111 - username@instance testConstraints> http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/dirlist.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/dirlist.md b/user_manual_1.3-incubating/examples/dirlist.md deleted file mode 100755 index cb419c5..0000000 --- a/user_manual_1.3-incubating/examples/dirlist.md +++ /dev/null @@ -1,43 +0,0 @@ ---- -title: File System Archive ---- - -This example shows how to use Accumulo to store a file system history. It has three classes: - - * Ingest.java - Recursively lists the files and directories under a given path, ingests their names and file info (not the file data!) into an Accumulo table, and indexes the file names in a separate table. - * QueryUtil.java - Provides utility methods for getting the info for a file, listing the contents of a directory, and performing single wild card searches on file or directory names. - * Viewer.java - Provides a GUI for browsing the file system information stored in Accumulo. - * FileCountMR.java - Runs MR over the file system information and writes out counts to an Accumulo table. - * FileCount.java - Accomplishes the same thing as FileCountMR, but in a different way. Computes recursive counts and stores them back into table. - * StringArraySummation.java - Aggregates counts for the FileCountMR reducer. - -To begin, ingest some data with Ingest.java. - - $ ./bin/accumulo org.apache.accumulo.examples.dirlist.Ingest instance zookeepers username password direxample dirindex exampleVis /local/user1/workspace - -Note that running this example will create tables direxample and dirindex in Accumulo that you should delete when you have completed the example. -If you modify a file or add new files in the directory ingested (e.g. /local/user1/workspace), you can run Ingest again to add new information into the Accumulo tables. - -To browse the data ingested, use Viewer.java. Be sure to give the "username" user the authorizations to see the data. - - $ ./bin/accumulo org.apache.accumulo.examples.dirlist.Viewer instance zookeepers username password direxample exampleVis /local/user1/workspace - -To list the contents of specific directories, use QueryUtil.java. - - $ ./bin/accumulo org.apache.accumulo.examples.dirlist.QueryUtil instance zookeepers username password direxample exampleVis /local/user1 - $ ./bin/accumulo org.apache.accumulo.examples.dirlist.QueryUtil instance zookeepers username password direxample exampleVis /local/user1/workspace - -To perform searches on file or directory names, also use QueryUtil.java. Search terms must contain no more than one wild card and cannot contain "/". -*Note* these queries run on the _dirindex_ table instead of the direxample table. - - $ ./bin/accumulo org.apache.accumulo.examples.dirlist.QueryUtil instance zookeepers username password dirindex exampleVis filename -search - $ ./bin/accumulo org.apache.accumulo.examples.dirlist.QueryUtil instance zookeepers username password dirindex exampleVis 'filename*' -search - $ ./bin/accumulo org.apache.accumulo.examples.dirlist.QueryUtil instance zookeepers username password dirindex exampleVis '*jar' -search - $ ./bin/accumulo org.apache.accumulo.examples.dirlist.QueryUtil instance zookeepers username password dirindex exampleVis filename*jar -search - -To count the number of direct children (directories and files) and descendants (children and children's descendents, directories and files), run the FileCountMR over the direxample table. -The results can be written back to the same table. - - $ ./bin/tool.sh lib/accumulo-examples-*[^c].jar org.apache.accumulo.examples.dirlist.FileCountMR instance zookeepers username password direxample direxample exampleVis exampleVis - -Alternatively, you can also run FileCount.java. http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/filter.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/filter.md b/user_manual_1.3-incubating/examples/filter.md deleted file mode 100755 index 08b297c..0000000 --- a/user_manual_1.3-incubating/examples/filter.md +++ /dev/null @@ -1,91 +0,0 @@ ---- -title: Filter Example ---- - -This is a simple filter example. It uses the AgeOffFilter that is provided as -part of the core package org.apache.accumulo.core.iterators.filter. Filters are used by -the FilteringIterator to select desired key/value pairs (or weed out undesired -ones). Filters implement the org.apache.accumulo.core.iterators.iterators.filter.Filter interface which -contains a method accept(Key k, Value v). This method returns true if the key, -value pair are to be delivered and false if they are to be ignored. - - username@instance> createtable filtertest - username@instance filtertest> setiter -t filtertest -scan -p 10 -n myfilter -filter - FilteringIterator uses Filters to accept or reject key/value pairs - ----------> entering options: <filterPriorityNumber> <ageoff|regex|filterClass> - ----------> set org.apache.accumulo.core.iterators.FilteringIterator option (<name> <value>, hit enter to skip): 0 ageoff - ----------> set org.apache.accumulo.core.iterators.FilteringIterator option (<name> <value>, hit enter to skip): - AgeOffFilter removes entries with timestamps more than <ttl> milliseconds old - ----------> set org.apache.accumulo.core.iterators.filter.AgeOffFilter parameter currentTime, if set, use the given value as the absolute time in milliseconds as the current time of day: - ----------> set org.apache.accumulo.core.iterators.filter.AgeOffFilter parameter ttl, time to live (milliseconds): 30000 - username@instance filtertest> - - username@instance filtertest> scan - username@instance filtertest> insert foo a b c - username@instance filtertest> scan - foo a:b [] c - -... wait 30 seconds ... - - username@instance filtertest> scan - username@instance filtertest> - -Note the absence of the entry inserted more than 30 seconds ago. Since the -scope was set to "scan", this means the entry is still in Accumulo, but is -being filtered out at query time. To delete entries from Accumulo based on -the ages of their timestamps, AgeOffFilters should be set up for the "minc" -and "majc" scopes, as well. - -To force an ageoff in the persisted data, after setting up the ageoff iterator -on the "minc" and "majc" scopes you can flush and compact your table. This will -happen automatically as a background operation on any table that is being -actively written to, but these are the commands to force compaction: - - username@instance filtertest> setiter -t filtertest -scan -minc -majc -p 10 -n myfilter -filter - FilteringIterator uses Filters to accept or reject key/value pairs - ----------> entering options: <filterPriorityNumber> <ageoff|regex|filterClass> - ----------> set org.apache.accumulo.core.iterators.FilteringIterator option (<name> <value>, hit enter to skip): 0 ageoff - ----------> set org.apache.accumulo.core.iterators.FilteringIterator option (<name> <value>, hit enter to skip): - AgeOffFilter removes entries with timestamps more than <ttl> milliseconds old - ----------> set org.apache.accumulo.core.iterators.filter.AgeOffFilter parameter currentTime, if set, use the given value as the absolute time in milliseconds as the current time of day: - ----------> set org.apache.accumulo.core.iterators.filter.AgeOffFilter parameter ttl, time to live (milliseconds): 30000 - username@instance filtertest> - - username@instance filtertest> flush -t filtertest - 08 11:13:55,745 [shell.Shell] INFO : Flush of table filtertest initiated... - username@instance filtertest> compact -t filtertest - 08 11:14:10,800 [shell.Shell] INFO : Compaction of table filtertest scheduled for 20110208111410EST - username@instance filtertest> - -After the compaction runs, the newly created files will not contain any data that should be aged off, and the -Accumulo garbage collector will remove the old files. - -To see the iterator settings for a table, use: - - username@instance filtertest> config -t filtertest -f iterator - ---------+------------------------------------------+---------------------------------------------------------- - SCOPE | NAME | VALUE - ---------+------------------------------------------+---------------------------------------------------------- - table | table.iterator.majc.myfilter .............. | 10,org.apache.accumulo.core.iterators.FilteringIterator - table | table.iterator.majc.myfilter.opt.0 ........ | org.apache.accumulo.core.iterators.filter.AgeOffFilter - table | table.iterator.majc.myfilter.opt.0.ttl .... | 30000 - table | table.iterator.majc.vers .................. | 20,org.apache.accumulo.core.iterators.VersioningIterator - table | table.iterator.majc.vers.opt.maxVersions .. | 1 - table | table.iterator.minc.myfilter .............. | 10,org.apache.accumulo.core.iterators.FilteringIterator - table | table.iterator.minc.myfilter.opt.0 ........ | org.apache.accumulo.core.iterators.filter.AgeOffFilter - table | table.iterator.minc.myfilter.opt.0.ttl .... | 30000 - table | table.iterator.minc.vers .................. | 20,org.apache.accumulo.core.iterators.VersioningIterator - table | table.iterator.minc.vers.opt.maxVersions .. | 1 - table | table.iterator.scan.myfilter .............. | 10,org.apache.accumulo.core.iterators.FilteringIterator - table | table.iterator.scan.myfilter.opt.0 ........ | org.apache.accumulo.core.iterators.filter.AgeOffFilter - table | table.iterator.scan.myfilter.opt.0.ttl .... | 30000 - table | table.iterator.scan.vers .................. | 20,org.apache.accumulo.core.iterators.VersioningIterator - table | table.iterator.scan.vers.opt.maxVersions .. | 1 - ---------+------------------------------------------+---------------------------------------------------------- - username@instance filtertest> - -If you would like to apply multiple filters, this can be done using a single -iterator. Just continue adding entries during the -"set org.apache.accumulo.core.iterators.FilteringIterator option" step. -Make sure to order the filterPriorityNumbers in the order you would like -the filters to be applied. http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/helloworld.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/helloworld.md b/user_manual_1.3-incubating/examples/helloworld.md deleted file mode 100755 index bff4977..0000000 --- a/user_manual_1.3-incubating/examples/helloworld.md +++ /dev/null @@ -1,38 +0,0 @@ ---- -title: Hello World Example ---- - -This tutorial uses the following Java classes, which can be found in org.apache.accumulo.examples.helloworld in the accumulo-examples module: - - * InsertWithBatchWriter.java - Inserts 10K rows (50K entries) into accumulo with each row having 5 entries - * InsertWithOutputFormat.java - Example of inserting data in MapReduce - * ReadData.java - Reads all data between two rows - -Log into the accumulo shell: - - $ ./bin/accumulo shell -u username -p password - -Create a table called 'hellotable': - - username@instance> createtable hellotable - -Launch a Java program that inserts data with a BatchWriter: - - $ ./bin/accumulo org.apache.accumulo.examples.helloworld.InsertWithBatchWriter instance zookeepers hellotable username password - -Alternatively, the same data can be inserted using MapReduce writers: - - $ ./bin/accumulo org.apache.accumulo.examples.helloworld.InsertWithOutputFormat instance zookeepers hellotable username password - -On the accumulo status page at the URL below (where 'master' is replaced with the name or IP of your accumulo master), you should see 50K entries - - http://master:50095/ - -To view the entries, use the shell to scan the table: - - username@instance> table hellotable - username@instance hellotable> scan - -You can also use a Java class to scan the table: - - $ ./bin/accumulo org.apache.accumulo.examples.helloworld.ReadData instance zookeepers hellotable username password row_0 row_1001 http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/index.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/index.md b/user_manual_1.3-incubating/examples/index.md deleted file mode 100644 index eaa3e4b..0000000 --- a/user_manual_1.3-incubating/examples/index.md +++ /dev/null @@ -1,42 +0,0 @@ ---- -title: Examples ---- - -Each README in the examples directory highlights the use of particular features of Apache Accumulo. - -Before running any of the examples, the following steps must be performed. - -1. Install and run Accumulo via the instructions found in $ACCUMULO_HOME/README. -Remember the instance name. It will be referred to as "instance" throughout the examples. -A comma-separated list of zookeeper servers will be referred to as "zookeepers". - -2. Create an Accumulo user (see the [user manual][1]), or use the root user. -The Accumulo user name will be referred to as "username" with password "password" throughout the examples. - -In all commands, you will need to replace "instance", "zookeepers", "username", and "password" with the values you set for your Accumulo instance. - -Commands intended to be run in bash are prefixed by '$'. These are always assumed to be run from the $ACCUMULO_HOME directory. - -Commands intended to be run in the Accumulo shell are prefixed by '>'. - -[1]: {{ site.baseurl }}/user_manual_1.3-incubating/Accumulo_Shell.html#User_Administration -[aggregation](aggregation.html) - -[batch](batch.html) - -[bloom](bloom.html) - -[bulkIngest](bulkIngest.html) - -[constraints](constraints.html) - -[dirlist](dirlist.html) - -[filter](filter.html) - -[helloworld](helloworld.html) - -[mapred](mapred.html) - -[shard](shard.html) - http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/mapred.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/mapred.md b/user_manual_1.3-incubating/examples/mapred.md deleted file mode 100755 index 975b5a0..0000000 --- a/user_manual_1.3-incubating/examples/mapred.md +++ /dev/null @@ -1,71 +0,0 @@ ---- -title: MapReduce Example ---- - -This example uses mapreduce and accumulo to compute word counts for a set of -documents. This is accomplished using a map-only mapreduce job and a -accumulo table with aggregators. - -To run this example you will need a directory in HDFS containing text files. -The accumulo readme will be used to show how to run this example. - - $ hadoop fs -copyFromLocal $ACCUMULO_HOME/README /user/username/wc/Accumulo.README - $ hadoop fs -ls /user/username/wc - Found 1 items - -rw-r--r-- 2 username supergroup 9359 2009-07-15 17:54 /user/username/wc/Accumulo.README - -The first part of running this example is to create a table with aggregation -for the column family count. - - $ ./bin/accumulo shell -u username -p password - Shell - Apache Accumulo Interactive Shell - - version: 1.3.x-incubating - - instance name: instance - - instance id: 00000000-0000-0000-0000-000000000000 - - - - type 'help' for a list of available commands - - - username@instance> createtable wordCount -a count=org.apache.accumulo.core.iterators.aggregation.StringSummation - username@instance wordCount> quit - -After creating the table, run the word count map reduce job. - - [user1@instance accumulo]$ bin/tool.sh lib/accumulo-examples-*[^c].jar org.apache.accumulo.examples.mapreduce.WordCount instance zookeepers /user/user1/wc wordCount -u username -p password - - 11/02/07 18:20:11 INFO input.FileInputFormat: Total input paths to process : 1 - 11/02/07 18:20:12 INFO mapred.JobClient: Running job: job_201102071740_0003 - 11/02/07 18:20:13 INFO mapred.JobClient: map 0% reduce 0% - 11/02/07 18:20:20 INFO mapred.JobClient: map 100% reduce 0% - 11/02/07 18:20:22 INFO mapred.JobClient: Job complete: job_201102071740_0003 - 11/02/07 18:20:22 INFO mapred.JobClient: Counters: 6 - 11/02/07 18:20:22 INFO mapred.JobClient: Job Counters - 11/02/07 18:20:22 INFO mapred.JobClient: Launched map tasks=1 - 11/02/07 18:20:22 INFO mapred.JobClient: Data-local map tasks=1 - 11/02/07 18:20:22 INFO mapred.JobClient: FileSystemCounters - 11/02/07 18:20:22 INFO mapred.JobClient: HDFS_BYTES_READ=10487 - 11/02/07 18:20:22 INFO mapred.JobClient: Map-Reduce Framework - 11/02/07 18:20:22 INFO mapred.JobClient: Map input records=255 - 11/02/07 18:20:22 INFO mapred.JobClient: Spilled Records=0 - 11/02/07 18:20:22 INFO mapred.JobClient: Map output records=1452 - -After the map reduce job completes, query the accumulo table to see word -counts. - - $ ./bin/accumulo shell -u username -p password - username@instance> table wordCount - username@instance wordCount> scan -b the - the count:20080906 [] 75 - their count:20080906 [] 2 - them count:20080906 [] 1 - then count:20080906 [] 1 - there count:20080906 [] 1 - these count:20080906 [] 3 - this count:20080906 [] 6 - through count:20080906 [] 1 - time count:20080906 [] 3 - time. count:20080906 [] 1 - to count:20080906 [] 27 - total count:20080906 [] 1 - tserver, count:20080906 [] 1 - tserver.compaction.major.concurrent.max count:20080906 [] 1 - ... http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/examples/shard.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/examples/shard.md b/user_manual_1.3-incubating/examples/shard.md deleted file mode 100755 index 0e0668a..0000000 --- a/user_manual_1.3-incubating/examples/shard.md +++ /dev/null @@ -1,52 +0,0 @@ ---- -title: Shard Example ---- - -Accumulo has an iterator called the intersecting iterator which supports querying a term index that is partitioned by -document, or "sharded". This example shows how to use the intersecting iterator through these four programs: - - * Index.java - Indexes a set of text files into an Accumulo table - * Query.java - Finds documents containing a given set of terms. - * Reverse.java - Reads the index table and writes a map of documents to terms into another table. - * ContinuousQuery.java Uses the table populated by Reverse.java to select N random terms per document. Then it continuously and randomly queries those terms. - -To run these example programs, create two tables like below. - - username@instance> createtable shard - username@instance shard> createtable doc2term - -After creating the tables, index some files. The following command indexes all of the java files in the Accumulo source code. - - $ cd /local/user1/workspace/accumulo/ - $ find src -name "*.java" | xargs ./bin/accumulo org.apache.accumulo.examples.shard.Index instance zookeepers shard username password 30 - -The following command queries the index to find all files containing 'foo' and 'bar'. - - $ cd $ACCUMULO_HOME - $ ./bin/accumulo org.apache.accumulo.examples.shard.Query instance zookeepers shard username password foo bar - /local/user1/workspace/accumulo/src/core/src/test/java/accumulo/core/security/ColumnVisibilityTest.java - /local/user1/workspace/accumulo/src/core/src/test/java/accumulo/core/client/mock/MockConnectorTest.java - /local/user1/workspace/accumulo/src/core/src/test/java/accumulo/core/security/VisibilityEvaluatorTest.java - /local/user1/workspace/accumulo/src/server/src/main/java/accumulo/server/test/functional/RowDeleteTest.java - /local/user1/workspace/accumulo/src/server/src/test/java/accumulo/server/logger/TestLogWriter.java - /local/user1/workspace/accumulo/src/server/src/main/java/accumulo/server/test/functional/DeleteEverythingTest.java - /local/user1/workspace/accumulo/src/core/src/test/java/accumulo/core/data/KeyExtentTest.java - /local/user1/workspace/accumulo/src/server/src/test/java/accumulo/server/constraints/MetadataConstraintsTest.java - /local/user1/workspace/accumulo/src/core/src/test/java/accumulo/core/iterators/WholeRowIteratorTest.java - /local/user1/workspace/accumulo/src/server/src/test/java/accumulo/server/util/DefaultMapTest.java - /local/user1/workspace/accumulo/src/server/src/test/java/accumulo/server/tabletserver/InMemoryMapTest.java - -Inorder to run ContinuousQuery, we need to run Reverse.java to populate doc2term - - $ ./bin/accumulo org.apache.accumulo.examples.shard.Reverse instance zookeepers shard doc2term username password - -Below ContinuousQuery is run using 5 terms. So it selects 5 random terms from each document, then it continually randomly selects one set of 5 terms and queries. It prints the number of matching documents and the time in seconds. - - $ ./bin/accumulo org.apache.accumulo.examples.shard.ContinuousQuery instance zookeepers shard doc2term username password 5 - [public, core, class, binarycomparable, b] 2 0.081 - [wordtodelete, unindexdocument, doctablename, putdelete, insert] 1 0.041 - [import, columnvisibilityinterpreterfactory, illegalstateexception, cv, columnvisibility] 1 0.049 - [getpackage, testversion, util, version, 55] 1 0.048 - [for, static, println, public, the] 55 0.211 - [sleeptime, wrappingiterator, options, long, utilwaitthread] 1 0.057 - [string, public, long, 0, wait] 12 0.132 http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/failure_handling.png ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/failure_handling.png b/user_manual_1.3-incubating/failure_handling.png deleted file mode 100644 index 90b9f0f..0000000 Binary files a/user_manual_1.3-incubating/failure_handling.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/img1.png ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/img1.png b/user_manual_1.3-incubating/img1.png deleted file mode 100644 index 8a5846c..0000000 Binary files a/user_manual_1.3-incubating/img1.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/img2.png ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/img2.png b/user_manual_1.3-incubating/img2.png deleted file mode 100644 index cbfe2b3..0000000 Binary files a/user_manual_1.3-incubating/img2.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/img3.png ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/img3.png b/user_manual_1.3-incubating/img3.png deleted file mode 100644 index 3b6f1f2..0000000 Binary files a/user_manual_1.3-incubating/img3.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/img4.png ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/img4.png b/user_manual_1.3-incubating/img4.png deleted file mode 100644 index 5b0ceb2..0000000 Binary files a/user_manual_1.3-incubating/img4.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/img5.png ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/img5.png b/user_manual_1.3-incubating/img5.png deleted file mode 100644 index 83d8955..0000000 Binary files a/user_manual_1.3-incubating/img5.png and /dev/null differ http://git-wip-us.apache.org/repos/asf/accumulo/blob/9a50bd13/user_manual_1.3-incubating/index.md ---------------------------------------------------------------------- diff --git a/user_manual_1.3-incubating/index.md b/user_manual_1.3-incubating/index.md deleted file mode 100644 index e40cda8..0000000 --- a/user_manual_1.3-incubating/index.md +++ /dev/null @@ -1,49 +0,0 @@ ---- -title: "User Manual: index" ---- - -** Next:** [Contents][2] ** [Contents][2]** - - - -## Apache Accumulo User Manual -Version 1.3 - - - - -* * * - -<a id="CHILD_LINKS"></a> - -* [Contents][2] -* [Introduction][6] -* [Accumulo Design][7] -* [Accumulo Shell][8] -* [Writing Accumulo Clients][9] -* [Table Configuration][10] -* [Table Design][11] -* [High-Speed Ingest][12] -* [Analytics][13] -* [Security][14] -* [Administration][15] -* [Shell Commands][16] - - - - -* * * - -[2]: Contents.html -[6]: Introduction.html -[7]: Accumulo_Design.html -[8]: Accumulo_Shell.html -[9]: Writing_Accumulo_Clients.html -[10]: Table_Configuration.html -[11]: Table_Design.html -[12]: High_Speed_Ingest.html -[13]: Analytics.html -[14]: Security.html -[15]: Administration.html -[16]: Shell_Commands.html -