Author: zblanco
Date: Wed Aug 26 14:15:27 2015
New Revision: 1697940
URL: http://svn.apache.org/r1697940
Log:
Fix numerous typos throughout the user-guide
Modified:
knox/trunk/books/0.7.0/book_client-details.md
knox/trunk/books/0.7.0/book_gateway-details.md
knox/trunk/books/0.7.0/book_knox-samples.md
knox/trunk/books/0.7.0/book_troubleshooting.md
knox/trunk/books/0.7.0/config.md
knox/trunk/books/0.7.0/config_advanced_ldap.md
knox/trunk/books/0.7.0/config_authn.md
knox/trunk/books/0.7.0/config_id_assertion.md
knox/trunk/books/0.7.0/config_ldap_group_lookup.md
knox/trunk/books/0.7.0/dev-guide/book.md
knox/trunk/books/0.7.0/knox_cli.md
knox/trunk/books/0.7.0/quick_start.md
knox/trunk/books/0.7.0/service_hbase.md
knox/trunk/books/0.7.0/service_hive.md
knox/trunk/books/0.7.0/service_oozie.md
knox/trunk/books/0.7.0/service_webhdfs.md
Modified: knox/trunk/books/0.7.0/book_client-details.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/book_client-details.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/book_client-details.md (original)
+++ knox/trunk/books/0.7.0/book_client-details.md Wed Aug 26 14:15:27 2015
@@ -31,7 +31,7 @@ The list below outlines the general requ
* Support the notion of a SSO session for multiple Hadoop interactions
* Support the multiple authentication and federation token capabilities of the
Apache Knox Gateway
* Promote the use of REST APIs as the dominant remote client mechanism for
Hadoop services
-* Promote the the sense of Hadoop as a single unified product
+* Promote the sense of Hadoop as a single unified product
* Aligned with the Apache Knox Gateway's overall goals for security
The result is a very simple DSL ([Domain Specific
Language](http://en.wikipedia.org/wiki/Domain-specific_language)) of sorts that
is used via [Groovy](http://groovy.codehaus.org) scripts.
Modified: knox/trunk/books/0.7.0/book_gateway-details.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/book_gateway-details.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/book_gateway-details.md (original)
+++ knox/trunk/books/0.7.0/book_gateway-details.md Wed Aug 26 14:15:27 2015
@@ -21,7 +21,7 @@ This section describes the details of th
* How URLs are mapped between a gateway that services multiple Hadoop clusters
and the clusters themselves
* How the gateway is configured through gateway-site.xml and cluster specific
topology files
-* How to configure the various policy enfocement provider features such as
authentication, authorization, auditing, hostmapping, etc.
+* How to configure the various policy enforcement provider features such as
authentication, authorization, auditing, hostmapping, etc.
### URL Mapping ###
Modified: knox/trunk/books/0.7.0/book_knox-samples.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/book_knox-samples.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/book_knox-samples.md (original)
+++ knox/trunk/books/0.7.0/book_knox-samples.md Wed Aug 26 14:15:27 2015
@@ -59,7 +59,7 @@ The Knox samples can however be made to
For manually installed Knox instances, there is really no way for the
installer to know how to configure the topology file for you.
-Essentially, these steps are identical to the Amabari deployed instance except
that #3 should be replaced with the configuration of the ootb sandbox.xml to
point the configuration at the proper hosts and ports.
+Essentially, these steps are identical to the Ambari deployed instance except
that #3 should be replaced with the configuration of the ootb sandbox.xml to
point the configuration at the proper hosts and ports.
1. You need to have ssh access to the environment in order for the localhost
assumption within the samples to be valid.
2. The Knox Demo LDAP Server is started - you can start it from Ambari
Modified: knox/trunk/books/0.7.0/book_troubleshooting.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/book_troubleshooting.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/book_troubleshooting.md (original)
+++ knox/trunk/books/0.7.0/book_troubleshooting.md Wed Aug 26 14:15:27 2015
@@ -85,7 +85,7 @@ If the gateway cannot contact one of the
at
org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784)
at
org.apache.hadoop.gateway.dispatch.HttpClientDispatch.executeRequest(HttpClientDispatch.java:99)
-The the resulting behavior on the client will differ by client.
+The resulting behavior on the client will differ by client.
For the client DSL executing the
{GATEWAY_HOME}/samples/ExampleWebHdfsLs.groovy the output will look look like
this.
Caught: org.apache.hadoop.gateway.shell.HadoopException:
org.apache.hadoop.gateway.shell.ErrorResponse: HTTP/1.1 500 Server Error
@@ -120,7 +120,7 @@ When Knox is configured to accept reques
the following error is returned
curl: (52) Empty reply from server
-This is the default behavior for Jetty SSL listener. While the credentials to
the default authentication provider continue to be username and password, we do
not want to encourage sending these in clear text. Since prememptively sending
BASIC credentials is a common pattern with REST APIs it would be unwise to
redirect to a HTTPS listener thus allowing clear text passwords.
+This is the default behavior for Jetty SSL listener. While the credentials to
the default authentication provider continue to be username and password, we do
not want to encourage sending these in clear text. Since preemptively sending
BASIC credentials is a common pattern with REST APIs it would be unwise to
redirect to a HTTPS listener thus allowing clear text passwords.
To resolve this issue, we have two options:
@@ -167,7 +167,7 @@ The client will likely see something alo
#### Using ldapsearch to verify ldap connectivtiy and credentials
-If your authentication to knox fails and you believe your are using correct
creedentilas, you could try to verify the connectivity and credentials usong
ldapsearch, assuming you are using ldap directory for authentication.
+If your authentication to knox fails and you believe your are using correct
creedentilas, you could try to verify the connectivity and credentials using
ldapsearch, assuming you are using ldap directory for authentication.
Assuming you are using the default values that came out of box with knox, your
ldapsearch command would be like the following
@@ -274,7 +274,7 @@ If the client hangs after emitting the l
Status : {...}
Creating table 'test_table'...
-HBase and Starget can be restred using the following commands on the Hadoop
Sandbox VM.
+HBase and Stargate can be restred using the following commands on the Hadoop
Sandbox VM.
You will need to ssh into the VM in order to run these commands.
sudo -u hbase /usr/lib/hbase/bin/hbase-daemon.sh stop master
@@ -285,7 +285,7 @@ You will need to ssh into the VM in orde
### SSL Certificate Issues ###
Clients that do not trust the certificate presented by the server will behave
in different ways.
-A browser will typically warn you of the inability to trust the receieved
certificate and give you an opportunity to add an exception for the particular
certificate.
+A browser will typically warn you of the inability to trust the received
certificate and give you an opportunity to add an exception for the particular
certificate.
Curl will present you with the follow message and instructions for turning of
certificate verification:
curl performs SSL certificate verification by default, using a "bundle"
Modified: knox/trunk/books/0.7.0/config.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/config.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/config.md (original)
+++ knox/trunk/books/0.7.0/config.md Wed Aug 26 14:15:27 2015
@@ -151,7 +151,7 @@ The general outline of a provider elemen
/topology/gateway/provider/role
: Defines the role of a particular provider.
-There are a number of pre-defined roles used by out-of-the-box provider
plugins for the gateay.
+There are a number of pre-defined roles used by out-of-the-box provider
plugins for the gateway.
These roles are: authentication, identity-assertion, authentication, rewrite
and hostmap
/topology/gateway/provider/name
@@ -217,7 +217,7 @@ The basic structure is shown below.
...
</topology>
-This mapping is required because the Hadoop servies running within the cluster
are unaware that they are being accessed from outside the cluster.
+This mapping is required because the Hadoop services running within the
cluster are unaware that they are being accessed from outside the cluster.
Therefore URLs returned as part of REST API responses will typically contain
internal host names.
Since clients outside the cluster will be unable to resolve those host name
they must be mapped to external host names.
@@ -337,7 +337,7 @@ Do not assume that the encryption if suf
A specific user should be created to run the gateway this user will be the
only user with permissions for the persisted master file.
-See the Knox CLI section for descriptions of the command line utilties related
to the master secret.
+See the Knox CLI section for descriptions of the command line utilities
related to the master secret.
#### Management of Security Artifacts ####
@@ -374,7 +374,7 @@ By leveraging the algorithm described ab
2. Using an NFS mount as a central location for the artifacts would provide a
single source of truth without the need to replicate them over the network. Of
course, NFS mounts have their own challenges.
3. Using the KnoxCLI to create and manage the security artifacts.
-See the Knox CLI section for descriptions of the command line utilties related
to the security artifact management.
+See the Knox CLI section for descriptions of the command line utilities
related to the security artifact management.
#### Keystores ####
In order to provide your own certificate for use by the gateway, you will need
to either import an existing key pair into a Java keystore or generate a
self-signed cert using the Java keytool.
@@ -412,13 +412,13 @@ The following will allow you to provisio
keytool -genkey -keyalg RSA -alias gateway-identity -keystore gateway.jks \
-storepass {master-secret} -validity 360 -keysize 2048
-Keytool will prompt you for a number of elements used will comprise the
distiniguished name (DN) within your certificate.
+Keytool will prompt you for a number of elements used will comprise the
distinguished name (DN) within your certificate.
*NOTE:* When it prompts you for your First and Last name be sure to type in
the hostname of the machine that your gateway instance will be running on. This
is used by clients during hostname verification to ensure that the presented
certificate matches the hostname that was used in the URL for the connection -
so they need to match.
*NOTE:* When it prompts for the key password just press enter to ensure that
it is the same as the keystore password. Which, as was described earlier, must
match the master secret for the gateway instance. Alternatively, you can set it
to another passphrase - take note of it and set the gateway-identity-passphrase
alias to that passphrase using the Knox CLI.
-See the Knox CLI section for descriptions of the command line utilties related
to the management of the keystores.
+See the Knox CLI section for descriptions of the command line utilities
related to the management of the keystores.
##### Using a CA Signed Key Pair #####
For certain deployments a certificate key pair that is signed by a trusted
certificate authority is required. There are a number of different ways in
which these certificates are acquired and can be converted and imported into
the Apache Knox keystore.
@@ -458,7 +458,7 @@ The credential stores in Knox use the JC
Keytool may be used to create credential stores but the Knox CLI section
details how to create aliases. These aliases are managed within credential
stores which are created by the CLI as needed. The simplest approach is to
create the gateway-identity-passpharse alias with the Knox CLI. This will
create the credential store if it doesn't already exist and add the key
passphrase.
-See the Knox CLI section for descriptions of the command line utilties related
to the management of the credential stores.
+See the Knox CLI section for descriptions of the command line utilities
related to the management of the credential stores.
##### Provisioning of Keystores #####
Once you have created these keystores you must move them into place for the
gateway to discover them and use them to represent its identity for SSL
connections. This is done by copying the keystores to the
`{GATEWAY_HOME}/data/security/keystores` directory for your gateway install.
Modified: knox/trunk/books/0.7.0/config_advanced_ldap.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/config_advanced_ldap.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/config_advanced_ldap.md (original)
+++ knox/trunk/books/0.7.0/config_advanced_ldap.md Wed Aug 26 14:15:27 2015
@@ -161,7 +161,7 @@ The configuration that you would use cou
<!-- search base used to search for user bind DN.
Defaults to the value of main.ldapRealm.searchBase.
If main.ldapRealm.userSearchAttributeName is defined,
- vlaue for main.ldapRealm.searchBase or
main.ldapRealm.userSearchBase
+ value for main.ldapRealm.searchBase or
main.ldapRealm.userSearchBase
should be defined -->
<param>
<name>main.ldapRealm.userSearchBase</name>
@@ -171,7 +171,7 @@ The configuration that you would use cou
<!-- search base used to search for groups.
Defaults to the value of main.ldapRealm.searchBase.
If value of main.ldapRealm.authorizationEnabled is true,
- vlaue for main.ldapRealm.searchBase or
main.ldapRealm.groupSearchBase should be defined -->
+ value for main.ldapRealm.searchBase or
main.ldapRealm.groupSearchBase should be defined -->
<param>
<name>main.ldapRealm.groupSearchBase</name>
<value>dc=hadoop,dc=apache,dc=org</value>
@@ -179,7 +179,7 @@ The configuration that you would use cou
<!-- optional, default value: groupOfNames
Objectclass to identify group entries in ldap, used to build
search
- filter to search for group entires -->
+ filter to search for group entries -->
<param>
<name>main.ldapRealm.groupObjectClass</name>
<value>groupOfNames</value>
@@ -222,7 +222,7 @@ The configuration that you would use cou
The value for this could have one of the following 2 formats
-plantextpassword
+plaintextpassword
${ALIAS=ldcSystemPassword}
The first format specifies the password in plain text in the provider
configuration.
Modified: knox/trunk/books/0.7.0/config_authn.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/config_authn.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/config_authn.md (original)
+++ knox/trunk/books/0.7.0/config_authn.md Wed Aug 26 14:15:27 2015
@@ -100,7 +100,7 @@ This section discusses the LDAP configur
**main.ldapRealm.userDnTemplate** - in order to bind a simple username to an
LDAP server that generally requires a full distinguished name (DN), we must
provide the template into which the simple username will be inserted. This
template allows for the creation of a DN by injecting the simple username into
the common name (CN) portion of the DN. **This element will need to be
customized to reflect your deployment environment.** The template provided in
the sample is only an example and is valid only within the LDAP schema
distributed with Knox and is represented by the users.ldif file in the
{GATEWAY_HOME}/conf directory.
-**main.ldapRealm.contextFactory.url** - this element is the URL that
represents the host and port of LDAP server. It also includes the scheme of the
protocol to use. This may be either ldap or ldaps depending on whether you are
communicating with the LDAP over SSL (higly recommended). **This element will
need to be cusomized to reflect your deployment environment.**.
+**main.ldapRealm.contextFactory.url** - this element is the URL that
represents the host and port of LDAP server. It also includes the scheme of the
protocol to use. This may be either ldap or ldaps depending on whether you are
communicating with the LDAP over SSL (highly recommended). **This element will
need to be customized to reflect your deployment environment.**.
**main.ldapRealm.contextFactory.authenticationMechanism** - this element
indicates the type of authentication that should be performed against the LDAP
server. The current default value is `simple` which indicates a simple bind
operation. This element should not need to be modified and no mechanism other
than a simple bind has been tested for this particular release.
@@ -110,7 +110,7 @@ This section discusses the LDAP configur
You would use LDAP configuration as documented above to authenticate against
Active Directory as well.
-Some Active Directory specifc things to keep in mind:
+Some Active Directory specific things to keep in mind:
Typical AD main.ldapRealm.userDnTemplate value looks slightly different, such
as
cn={0},cn=users,DC=lab,DC=sample,dc=com
@@ -125,7 +125,7 @@ If your AD is configured to authenticate
In order to communicate with your LDAP server over SSL (again, highly
recommended), you will need to modify the topology file in a couple ways and
possibly provision some keying material.
1. **main.ldapRealm.contextFactory.url** must be changed to have the `ldaps`
protocol scheme and the port must be the SSL listener port on your LDAP server.
-2. Identity certificate (keypair) provisioned to LDAP server - your LDAP
server specific documentation should indicate what is requried for providing a
cert or keypair to represent the LDAP server identity to connecting clients.
+2. Identity certificate (keypair) provisioned to LDAP server - your LDAP
server specific documentation should indicate what is required for providing a
cert or keypair to represent the LDAP server identity to connecting clients.
3. Trusting the LDAP Server's public key - if the LDAP Server's identity
certificate is issued by a well known and trusted certificate authority and is
already represented in the JRE's cacerts truststore then you don't need to do
anything for trusting the LDAP server's cert. If, however, the cert is
selfsigned or issued by an untrusted authority you will need to either add it
to the cacerts keystore or to another truststore that you may direct Knox to
utilize through a system property.
#### Session Configuration ####
Modified: knox/trunk/books/0.7.0/config_id_assertion.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/config_id_assertion.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/config_id_assertion.md (original)
+++ knox/trunk/books/0.7.0/config_id_assertion.md Wed Aug 26 14:15:27 2015
@@ -33,7 +33,7 @@ The following configuration is required
<enabled>true</enabled>
</provider>
-This particular configuration indicates that the Default identity assertion
provider is enabled and that there are no principal mapping rules to apply to
identities flowing from the authentication in the gateway to the backend Hadoop
cluster services. The primary principal of the current subject will therefore
be asserted via a query paramter or as a form parameter - ie.
?user.name={primaryPrincipal}
+This particular configuration indicates that the Default identity assertion
provider is enabled and that there are no principal mapping rules to apply to
identities flowing from the authentication in the gateway to the backend Hadoop
cluster services. The primary principal of the current subject will therefore
be asserted via a query parameter or as a form parameter - ie.
?user.name={primaryPrincipal}
<provider>
<role>identity-assertion</role>
Modified: knox/trunk/books/0.7.0/config_ldap_group_lookup.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/config_ldap_group_lookup.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/config_ldap_group_lookup.md (original)
+++ knox/trunk/books/0.7.0/config_ldap_group_lookup.md Wed Aug 26 14:15:27 2015
@@ -33,7 +33,7 @@ Please see below a sample Shiro configur
<!--
session timeout in minutes, this is really idle timeout,
defaults to 30mins, if the property value is not defined,,
- current client authentication would expire if client idles
contiuosly for more than this value
+ current client authentication would expire if client idles
continuosly for more than this value
-->
<!-- defaults to: 30 minutes
<param>
@@ -168,12 +168,12 @@ java -jar bin/ldap.jar conf
java -Dsandbox.ldcSystemPassword=guest-password -jar bin/gateway.jar
-persist-master
Following call to WebHDFS should report HTTP/1.1 401 Unauthorized
-As guest is not a member of group "analyst", authorization prvoider states
user should be member of group "analyst"
+As guest is not a member of group "analyst", authorization provider states
user should be member of group "analyst"
curl -i -v -k -u guest:guest-password -X GET
https://localhost:8443/gateway/sandbox/webhdfs/v1?op=GETHOMEDIRECTORY
Following call to WebHDFS should report: {"Path":"/user/sam"}
-As sam is a member of group "analyst", authorization prvoider states user
should be member of group "analyst"
+As sam is a member of group "analyst", authorization provider states user
should be member of group "analyst"
curl -i -v -k -u sam:sam-password -X GET
https://localhost:8443/gateway/sandbox/webhdfs/v1?op=GETHOMEDIRECTORY
@@ -192,12 +192,12 @@ java -jar bin/ldap.jar conf
java -Dsandbox.ldcSystemPassword=guest-password -jar bin/gateway.jar
-persist-master
Following call to WebHDFS should report HTTP/1.1 401 Unauthorized
-As guest is not a member of group "analyst", authorization prvoider states
user should be member of group "analyst"
+As guest is not a member of group "analyst", authorization provider states
user should be member of group "analyst"
curl -i -v -k -u guest:guest-password -X GET
https://localhost:8443/gateway/sandbox/webhdfs/v1?op=GETHOMEDIRECTORY
Following call to WebHDFS should report: {"Path":"/user/sam"}
-As sam is a member of group "analyst", authorization prvoider states user
should be member of group "analyst"
+As sam is a member of group "analyst", authorization provxider states user
should be member of group "analyst"
curl -i -v -k -u sam:sam-password -X GET
https://localhost:8443/gateway/sandbox/webhdfs/v1?op=GETHOMEDIRECTORY
@@ -214,15 +214,15 @@ cp templates/users.ldapdynamicgroups.ldi
java -jar bin/ldap.jar conf
java -Dsandbox.ldcSystemPassword=guest-password -jar bin/gateway.jar
-persist-master
-Please note that user.ldapdynamicgroups.ldif also loads ncessary schema to
create dynamic groups in Apache DS.
+Please note that user.ldapdynamicgroups.ldif also loads necessary schema to
create dynamic groups in Apache DS.
Following call to WebHDFS should report HTTP/1.1 401 Unauthorized
-As guest is not a member of dynamic group "directors", authorization prvoider
states user should be member of group "directors"
+As guest is not a member of dynamic group "directors", authorization provider
states user should be member of group "directors"
curl -i -v -k -u guest:guest-password -X GET
https://localhost:8443/gateway/sandbox/webhdfs/v1?op=GETHOMEDIRECTORY
Following call to WebHDFS should report: {"Path":"/user/bob"}
-As bob is a member of dynamic group "directors", authorization prvoider states
user should be member of group "directors"
+As bob is a member of dynamic group "directors", authorization provider states
user should be member of group "directors"
curl -i -v -k -u sam:sam-password -X GET
https://localhost:8443/gateway/sandbox/webhdfs/v1?op=GETHOMEDIRECTORY
Modified: knox/trunk/books/0.7.0/dev-guide/book.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/dev-guide/book.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/dev-guide/book.md (original)
+++ knox/trunk/books/0.7.0/dev-guide/book.md Wed Aug 26 14:15:27 2015
@@ -502,7 +502,7 @@ public void testDevGuideSample() throws
There are a number of extension points available in the gateway: services,
providers, rewrite steps and functions, etc.
All of these use the Java ServiceLoader mechanism for their discovery.
There are two ways to make these extensions available on the class path at
runtime.
-The first way to to add a new module to the project and have the extension
"built-in".
+The first way to add a new module to the project and have the extension
"built-in".
The second is to add the extension to the class path of the server after it is
installed.
Both mechanism are described in more detail below.
Modified: knox/trunk/books/0.7.0/knox_cli.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/knox_cli.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/knox_cli.md (original)
+++ knox/trunk/books/0.7.0/knox_cli.md Wed Aug 26 14:15:27 2015
@@ -26,7 +26,7 @@ The knoxcli.sh script is located in the
##### `bin/knoxcli.sh [--help]` #####
prints help for all commands
-#### Knox Verison Info ####
+#### Knox Version Info ####
##### `bin/knoxcli.sh version [--help]` #####
Displays Knox version information.
Modified: knox/trunk/books/0.7.0/quick_start.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/quick_start.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/quick_start.md (original)
+++ knox/trunk/books/0.7.0/quick_start.md Wed Aug 26 14:15:27 2015
@@ -195,7 +195,7 @@ To validate that see the sections for th
'https://localhost:8443/gateway/sandbox/webhdfs/v1/tmp/LICENSE?op=CREATE'
curl -i -k -u guest:guest-password -T LICENSE -X PUT \
- '{Value of Location header from response response above}'
+ '{Value of Location header from response above}'
#### Get a file in HDFS via Knox.
Modified: knox/trunk/books/0.7.0/service_hbase.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/service_hbase.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/service_hbase.md (original)
+++ knox/trunk/books/0.7.0/service_hbase.md Wed Aug 26 14:15:27 2015
@@ -29,7 +29,7 @@ See the HBase Stargate Setup section bel
#### HBase Examples ####
The examples below illustrate the set of basic operations with HBase instance
using Stargate REST API.
-Use following link to get more more details about HBase/Stargate API:
http://wiki.apache.org/hadoop/Hbase/Stargate.
+Use following link to get more details about HBase/Stargate API:
http://wiki.apache.org/hadoop/Hbase/Stargate.
Note: Some HBase examples may not work due to enabled [Access
Control](https://hbase.apache.org/book/hbase.accesscontrol.configuration.html).
User may not be granted for performing operations in samples. In order to check
if Access Control is configured in the HBase instance verify hbase-site.xml for
a presence of `org.apache.hadoop.hbase.security.access.AccessController` in
`hbase.coprocessor.master.classes` and `hbase.coprocessor.region.classes`
properties.
To grant the Read, Write, Create permissions to `guest` user execute the
following command:
@@ -277,7 +277,7 @@ After launching the shell, execute the f
* endTime(Long) - the upper bound for filtration by time.
* times(Long startTime, Long endTime) - the lower and upper bounds for
filtration by time.
* filter(String) - the filter XML definition.
- * maxVersions(Integer) - the the maximum number of versions to return.
+ * maxVersions(Integer) - the maximum number of versions to return.
* Response
* scannerId : String - the scanner ID of the created scanner. Consumes
body.
* Example
Modified: knox/trunk/books/0.7.0/service_hive.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/service_hive.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/service_hive.md (original)
+++ knox/trunk/books/0.7.0/service_hive.md Wed Aug 26 14:15:27 2015
@@ -63,7 +63,7 @@ By default the gateway is configured to
#### Hive Examples ####
-This guide provides detailed examples for how to to some basic interactions
with Hive via the Apache Knox Gateway.
+This guide provides detailed examples for how to do some basic interactions
with Hive via the Apache Knox Gateway.
##### Hive Setup #####
@@ -231,7 +231,7 @@ Each line from the file below will need
statement.close();
connection.close();
-Exampes use 'log.txt' with content:
+Examples use 'log.txt' with content:
2012-02-03 18:35:34 SampleClass6 [INFO] everything normal for id 577725851
2012-02-03 18:35:34 SampleClass4 [FATAL] system problem at id 1991281254
Modified: knox/trunk/books/0.7.0/service_oozie.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/service_oozie.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/service_oozie.md (original)
+++ knox/trunk/books/0.7.0/service_oozie.md Wed Aug 26 14:15:27 2015
@@ -21,7 +21,7 @@
Oozie is a Hadoop component provides complex job workflows to be submitted and
managed.
Please refer to the latest [Oozie
documentation](http://oozie.apache.org/docs/4.0.0/) for details.
-In order to make Oozie accessible via the gateway there are several important
Haddop configuration settings.
+In order to make Oozie accessible via the gateway there are several important
Hadoop configuration settings.
These all relate to the network endpoint exposed by various Hadoop services.
The HTTP endpoint at which Oozie is running can be found via the
oozie.base.url property in the oozie-site.xml file.
Modified: knox/trunk/books/0.7.0/service_webhdfs.md
URL:
http://svn.apache.org/viewvc/knox/trunk/books/0.7.0/service_webhdfs.md?rev=1697940&r1=1697939&r2=1697940&view=diff
==============================================================================
--- knox/trunk/books/0.7.0/service_webhdfs.md (original)
+++ knox/trunk/books/0.7.0/service_webhdfs.md Wed Aug 26 14:15:27 2015
@@ -215,7 +215,7 @@ Use can use cURL to directly invoke the
###### put() - Write a file into HDFS (CREATE)
* Request
- * text( String text ) - Text to upload to HDFS. Takes precidence over
file if both present.
+ * text( String text ) - Text to upload to HDFS. Takes precedence over
file if both present.
* file( String name ) - The name of a local file to upload to HDFS.
* to( String name ) - The fully qualified name to create in HDFS.
* Response