sijie closed pull request #2335: [documentation] Improve documentation on bare 
mental deployment
URL: https://github.com/apache/incubator-pulsar/pull/2335
 
 
   

This is a PR merged from a forked repository.
As GitHub hides the original diff on merge, it is displayed below for
the sake of provenance:

As this is a foreign pull request (from a fork), the diff is supplied
below (as it won't show otherwise due to GitHub magic):

diff --git a/deployment/terraform-ansible/aws/setup-disk.yaml 
b/deployment/terraform-ansible/aws/setup-disk.yaml
new file mode 100644
index 0000000000..e1360c0abe
--- /dev/null
+++ b/deployment/terraform-ansible/aws/setup-disk.yaml
@@ -0,0 +1,36 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+#
+
+- name: Disk setup
+  hosts: pulsar
+  connection: ssh
+  become: true
+  tasks:
+    - command: >
+        tuned-adm profile latency-performance
+    - name: Create and mount disks
+      mount:
+        path: "{{ item.path }}"
+        src: "{{ item.src }}"
+        fstype: xfs
+        opts: defaults,noatime,nodiscard
+        state: present
+      with_items:
+        - { path: "/mnt/journal", src: "/dev/nvme0n1" }
+        - { path: "/mnt/storage", src: "/dev/nvme1n1" }
diff --git a/deployment/terraform-ansible/deploy-pulsar.yaml 
b/deployment/terraform-ansible/deploy-pulsar.yaml
index 78dc3a17d6..d586a5e1e0 100644
--- a/deployment/terraform-ansible/deploy-pulsar.yaml
+++ b/deployment/terraform-ansible/deploy-pulsar.yaml
@@ -17,24 +17,6 @@
 # under the License.
 #
 
-- name: Disk setup
-  hosts: pulsar
-  connection: ssh
-  become: true
-  tasks:
-    - command: >
-        tuned-adm profile latency-performance
-    - name: Create and mount disks
-      mount:
-        path: "{{ item.path }}"
-        src: "{{ item.src }}"
-        fstype: xfs
-        opts: defaults,noatime,nodiscard
-        state: present
-      with_items:
-        - { path: "/mnt/journal", src: "/dev/nvme0n1" }
-        - { path: "/mnt/storage", src: "/dev/nvme1n1" }
-
 - name: Pulsar setup
   hosts: all
   connection: ssh
@@ -56,7 +38,7 @@
         zookeeper_servers: "{{ groups['zookeeper']|map('extract', hostvars, 
['ansible_default_ipv4', 'address'])|map('regex_replace', '(.*)', '\\1:2181') | 
join(',') }}"
         service_url: "pulsar://{{ hostvars[groups['pulsar'][0]].public_ip 
}}:6650/"
         http_url: "http://{{ hostvars[groups['pulsar'][0]].public_ip }}:8080/"
-        pulsar_version: "2.0.0-rc1-incubating"
+        pulsar_version: "2.1.0-incubating"
 
     - name: Download Pulsar binary package
       unarchive:
diff --git a/deployment/terraform-ansible/templates/bookkeeper.conf 
b/deployment/terraform-ansible/templates/bookkeeper.conf
index 2f5dd4ae59..9e7fcc9d7c 100644
--- a/deployment/terraform-ansible/templates/bookkeeper.conf
+++ b/deployment/terraform-ansible/templates/bookkeeper.conf
@@ -53,7 +53,7 @@ minUsableSizeForIndexFileCreation=1073741824
 # Configure a specific hostname or IP address that the bookie should use to 
advertise itself to
 # clients. If not set, bookie will advertised its own IP address or hostname, 
depending on the
 # listeningInterface and `seHostNameAsBookieID settings.
-advertisedAddress={{ hostvars[inventory_hostname].public_ip }}
+# advertisedAddress=
 
 # Whether the bookie allowed to use a loopback interface as its primary
 # interface(i.e. the interface it uses to establish its identity)?
diff --git a/site2/docs/deploy-aws.md b/site2/docs/deploy-aws.md
index 1ef3cc3db0..f827177457 100644
--- a/site2/docs/deploy-aws.md
+++ b/site2/docs/deploy-aws.md
@@ -41,6 +41,24 @@ $ cd incubator-pulsar/deployment/terraform-ansible/aws
 
 ## SSH setup
 
+> If you already have an SSH key and would like to use it, you skip generating 
the SSH keys and update `private_key_file` setting
+> in `ansible.cfg` file and `public_key_path` setting in `terraform.tfvars` 
file.
+>
+> For example, if you already had a private SSH key in `~/.ssh/pulsar_aws` and 
a public key in `~/.ssh/pulsar_aws.pub`,
+> you can do followings:
+>
+> 1. update `ansible.cfg` with following values:
+>
+> ```shell
+> private_key_file=~/.ssh/pulsar_aws
+> ```
+>
+> 2. update `terraform.tfvars` with following values:
+>
+> ```shell
+> public_key_path=~/.ssh/pulsar_aws.pub
+> ```
+
 In order to create the necessary AWS resources using Terraform, you'll need to 
create an SSH key. To create a private SSH key in `~/.ssh/id_rsa` and a public 
key in `~/.ssh/id_rsa.pub`:
 
 ```bash
@@ -133,6 +151,25 @@ At any point, you can destroy all AWS resources associated 
with your cluster usi
 $ terraform destroy
 ```
 
+## Setup Disks
+
+Before you run the Pulsar playbook, you want to mount the disks to the correct 
directories on those bookie nodes.
+Since different type of machines would have different disk layout, if you 
change the `instance_types` in your terraform
+config, you need to update the task defined in `setup-disk.yaml` file.
+
+To setup disks on bookie nodes, use this command:
+
+```bash
+$ ansible-playbook \
+  --user='ec2-user' \
+  --inventory=`which terraform-inventory` \
+  setup-disk.yaml
+```
+
+After running this command, the disks will be mounted under `/mnt/journal` as 
journal disk, and `/mnt/storage` as ledger disk.
+It is important to run this command only once! If you attempt to run this 
command again after you have run Pulsar playbook,
+it might be potentially erase your disks again and cause the bookies to fail 
to start up.
+
 ## Running the Pulsar playbook
 
 Once you've created the necessary AWS resources using Terraform, you can 
install and run Pulsar on the Terraform-created EC2 instances using Ansible. To 
do so, use this command:
@@ -150,7 +187,6 @@ If you've created a private SSH key at a location different 
from `~/.ssh/id_rsa`
 $ ansible-playbook \
   --user='ec2-user' \
   --inventory=`which terraform-inventory` \
-  --private-key="~/.ssh/some-non-default-key" \
   ../deploy-pulsar.yaml
 ```
 
diff --git a/site2/docs/deploy-bare-metal.md b/site2/docs/deploy-bare-metal.md
index c5ea8939f7..78ea557b6f 100644
--- a/site2/docs/deploy-bare-metal.md
+++ b/site2/docs/deploy-bare-metal.md
@@ -17,18 +17,23 @@ sidebar_label: Bare metal
 
 Deploying a Pulsar cluster involves doing the following (in order):
 
-* Deploying a [ZooKeeper](#deploying-a-zookeeper-cluster) cluster
+* Deploying a [ZooKeeper](#deploying-a-zookeeper-cluster) cluster (optional)
 * Initializing [cluster metadata](#initializing-cluster-metadata)
 * Deploying a [BookKeeper](#deploying-a-bookkeeper-cluster) cluster
 * Deploying one or more Pulsar [brokers](#deploying-pulsar-brokers)
 
+## Preparation
+
 ### Requirements
 
+> If you already have an existing zookeeper cluster and would like to reuse 
it, you don't need to prepare the machines
+> for running ZooKeeper.
+
 To run Pulsar on bare metal, you will need:
 
 * At least 6 Linux machines or VMs
   * 3 running [ZooKeeper](https://zookeeper.apache.org)
-  * 3 running a Pulsar broker and a 
[BookKeeper](https://bookkeeper.apache.org) bookie
+  * 3 running a Pulsar broker, and a 
[BookKeeper](https://bookkeeper.apache.org) bookie
 * A single [DNS](https://en.wikipedia.org/wiki/Domain_Name_System) name 
covering all of the Pulsar broker hosts
 
 Each machine in your cluster will need to have [Java 
8](http://www.oracle.com/technetwork/java/javase/downloads/index.html) or 
higher installed.
@@ -43,8 +48,12 @@ In this diagram, connecting clients need to be able to 
communicate with the Puls
 
 When deploying a Pulsar cluster, we have some basic recommendations that you 
should keep in mind when capacity planning.
 
+#### ZooKeeper
+
 For machines running ZooKeeper, we recommend using lighter-weight machines or 
VMs. Pulsar uses ZooKeeper only for periodic coordination- and 
configuration-related tasks, *not* for basic operations. If you're running 
Pulsar on [Amazon Web Services](https://aws.amazon.com/) (AWS), for example, a 
[t2.small](http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/t2-instances.html)
 instance would likely suffice.
 
+#### Bookies & Brokers
+
 For machines running a bookie and a Pulsar broker, we recommend using more 
powerful machines. For an AWS deployment, for example, 
[i3.4xlarge](https://aws.amazon.com/blogs/aws/now-available-i3-instances-for-demanding-io-intensive-applications/)
 instances may be appropriate. On those machines we also recommend:
 
 * Fast CPUs and 10Gbps 
[NIC](https://en.wikipedia.org/wiki/Network_interface_controller) (for Pulsar 
brokers)
@@ -83,8 +92,52 @@ Directory | Contains
 `lib` | The [JAR](https://en.wikipedia.org/wiki/JAR_(file_format)) files used 
by Pulsar.
 `logs` | Logs created by the installation.
 
+## Installing Builtin Connectors (optional)
+
+> Since release `2.1.0-incubating`, Pulsar releases a separate binary 
distribution, containing all the `builtin` connectors.
+> If you would like to enable those `builtin` connectors, you can follow the 
instructions as below; otherwise you can
+> skip this section for now.
+
+To get started using builtin connectors, you'll need to download the 
connectors tarball release on every broker node in
+one of the following ways:
+
+* by clicking the link below and downloading the release from an Apache mirror:
+
+  * <a href="pulsar:connector_release_url" download>Pulsar IO Connectors 
{{pulsar:version}} release</a>
+
+* from the Pulsar [downloads page](pulsar:download_page_url)
+* from the Pulsar [releases 
page](https://github.com/apache/incubator-pulsar/releases/latest)
+* using [wget](https://www.gnu.org/software/wget):
+
+  ```shell
+  $ wget pulsar:connector_release_url
+  ```
+
+Once the tarball is downloaded, in the pulsar directory, untar the 
io-connectors package and copy the connectors as `connectors`
+in the pulsar directory:
+
+```bash
+$ tar xvfz apache-pulsar-io-connectors-{{pulsar:version}}-bin.tar.gz
+
+// you will find a directory named 
`apache-pulsar-io-connectors-{{pulsar:version}}` in the pulsar directory
+// then copy the connectors
+
+$ mv apache-pulsar-io-connectors-{{pulsar:version}}/connectors connectors
+
+$ ls connectors
+pulsar-io-aerospike-{{pulsar.version}}.nar
+pulsar-io-cassandra-{{pulsar.version}}.nar 
+pulsar-io-kafka-{{pulsar.version}}.nar     
+pulsar-io-kinesis-{{pulsar.version}}.nar   
+pulsar-io-rabbitmq-{{pulsar.version}}.nar  
+pulsar-io-twitter-{{pulsar.version}}.nar
+...
+```
+
 ## Deploying a ZooKeeper cluster
 
+> If you already have an exsiting zookeeper cluster and would like to use it, 
you can skip this section.
+
 [ZooKeeper](https://zookeeper.apache.org) manages a variety of essential 
coordination- and configuration-related tasks for Pulsar. To deploy a Pulsar 
cluster you'll need to deploy ZooKeeper first (before all other components). We 
recommend deploying a 3-node ZooKeeper cluster. Pulsar does not make heavy use 
of ZooKeeper, so more lightweight machines or VMs should suffice for running 
ZooKeeper.
 
 To begin, add all ZooKeeper servers to the configuration specified in 
[`conf/zookeeper.conf`](reference-configuration.md#zookeeper) (in the Pulsar 
directory you created [above](#installing-the-pulsar-binary-package)). Here's 
an example:
@@ -155,6 +208,15 @@ 
zkServers=zk1.us-west.example.com:2181,zk2.us-west.example.com:2181,zk3.us-west.
 
 Once you've appropriately modified the `zkServers` parameter, you can provide 
any other configuration modifications you need. You can find a full listing of 
the available BookKeeper configuration parameters 
[here](reference-configuration.md#bookkeeper), although we would recommend 
consulting the [BookKeeper 
documentation](http://bookkeeper.apache.org/docs/latest/reference/config/) for 
a more in-depth guide.
 
+> ##### NOTES
+>
+> Since Pulsar 2.1.0 release, Pulsar introduces [stateful 
function](functions-state.md) for Pulsar Functions. If you would like to enable 
that feature,
+> you need to enable table service on BookKeeper by setting following setting 
in `conf/bookkeeper.conf` file.
+>
+> ```conf
+> 
extraServerComponents=org.apache.bookkeeper.stream.server.StreamStorageLifecycleComponent
+> ```
+
 Once you've applied the desired configuration in `conf/bookkeeper.conf`, you 
can start up a bookie on each of your BookKeeper hosts. You can start up each 
bookie either in the background, using 
[nohup](https://en.wikipedia.org/wiki/Nohup), or in the foreground.
 
 To start the bookie in the background, use the 
[`pulsar-daemon`](reference-cli-tools.md#pulsar-daemon) CLI tool:
@@ -169,7 +231,7 @@ To start the bookie in the foreground:
 $ bin/bookkeeper bookie
 ```
 
-You can verify that the bookie is working properly using the `bookiesanity` 
command for the [BookKeeper 
shell](http://localhost:4000/docs/latest/deployment/reference/CliTools#bookkeeper-shell):
+You can verify that a bookie is working properly by running the `bookiesanity` 
command for the [BookKeeper shell](reference-cli-tools.md#shell) on it:
 
 ```bash
 $ bin/bookkeeper shell bookiesanity
@@ -177,10 +239,22 @@ $ bin/bookkeeper shell bookiesanity
 
 This will create an ephemeral BookKeeper ledger on the local bookie, write a 
few entries, read them back, and finally delete the ledger.
 
+After you have started all the bookies, you can use `simpletest` command for 
[BookKeeper shell](reference-cli-tools.md#shell) on any bookie node, to
+verify all the bookies in the cluster are up running.
+
+```bash
+$ bin/bookkeeper shell simpletest --ensemble <num-bookies> --writeQuorum 
<num-bookies> --ackQuorum <num-bookies> --numEntries <num-entries>
+```
+
+This command will create a `num-bookies` sized ledger on the cluster, write a 
few entries, and finally delete the ledger.
+
+
 ## Deploying Pulsar brokers
 
 Pulsar brokers are the last thing you need to deploy in your Pulsar cluster. 
Brokers handle Pulsar messages and provide Pulsar's administrative interface. 
We recommend running **3 brokers**, one for each machine that's already running 
a BookKeeper bookie.
 
+### Configuring Brokers
+
 The most important element of broker configuration is ensuring that that each 
broker is aware of the ZooKeeper cluster that you've deployed. Make sure that 
the [`zookeeperServers`](reference-configuration.md#broker-zookeeperServers) 
and 
[`configurationStoreServers`](reference-configuration.md#broker-configurationStoreServers)
 parameters. In this case, since we only have 1 cluster and no configuration 
store setup, the `configurationStoreServers` will point to the same 
`zookeeperServers`.
 
 ```properties
@@ -194,6 +268,24 @@ You also need to specify the cluster name (matching the 
name that you provided w
 clusterName=pulsar-cluster-1
 ```
 
+### Enabling Pulsar Functions (optional)
+
+If you want to enable [Pulsar Functions](functions-overview.md), you can 
follow the instructions as below:
+
+1. Edit `conf/broker.conf` to enable function worker, by setting 
`functionsWorkerEnabled` to `true`.
+
+    ```conf
+    functionsWorkerEnabled=true
+    ```
+
+2. Edit `conf/functions_worker.yml` and set `pulsarFunctionsCluster` to the 
cluster name that you provided when [initializing the cluster's 
metadata](#initializing-cluster-metadata). 
+
+    ```conf
+    pulsarFunctionsCluster=pulsar-cluster-1
+    ```
+
+### Starting Brokers
+
 You can then provide any other configuration changes that you'd like in the 
[`conf/broker.conf`](reference-configuration.md#broker) file. Once you've 
decided on a configuration, you can start up the brokers for your Pulsar 
cluster. Like ZooKeeper and BookKeeper, brokers can be started either in the 
foreground or in the background, using nohup.
 
 You can start a broker in the foreground using the [`pulsar 
broker`](reference-cli-tools.md#pulsar-broker) command:
@@ -233,3 +325,32 @@ $ bin/pulsar-client produce \
 > You may need to use a different cluster name in the topic if you specified a 
 > cluster name different from `pulsar-cluster-1`.
 
 This will publish a single message to the Pulsar topic.
+
+## Running Functions
+
+> If you have [enabled](#enabling-pulsar-functions-optional) Pulsar Functions, 
you can also tryout pulsar functions now.
+
+Create a ExclamationFunction `exclamation`.
+
+```bash
+bin/pulsar-admin functions create \
+  --jar examples/api-examples.jar \
+  --className org.apache.pulsar.functions.api.examples.ExclamationFunction \
+  --inputs persistent://public/default/exclamation-input \
+  --output persistent://public/default/exclamation-output \
+  --tenant public \
+  --namespace default \
+  --name exclamation
+```
+
+Check if the function is running as expected by 
[triggering](functions-deploying.md#triggering-pulsar-functions) the function.
+
+```bash
+bin/pulsar-admin functions trigger --name exclamation --triggerValue "hello 
world"
+```
+
+You will see output as below:
+
+```shell
+hello world!
+```


 

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

Reply via email to