Github user HeartSaVioR commented on a diff in the pull request:
https://github.com/apache/storm/pull/1666#discussion_r77107492
--- Diff: docs/storm-sql.md ---
@@ -104,21 +104,22 @@ Let's say there is a Kafka stream that represents the
transactions of orders. Ea
The user can specify the following SQL statements in the SQL file:
```
-CREATE EXTERNAL TABLE ORDERS (ID INT PRIMARY KEY, UNIT_PRICE INT, QUANTITY
INT) LOCATION 'kafka://localhost:2181/brokers?topic=orders' TBLPROPERTIES
'{"producer":{"bootstrap.servers":"localhost:9092","acks":"1","key.serializer":"org.apache.org.apache.storm.kafka.IntSerializer","value.serializer":"org.apache.org.apache.storm.kafka.ByteBufferSerializer"}}'
+CREATE EXTERNAL TABLE ORDERS (ID INT PRIMARY KEY, UNIT_PRICE INT, QUANTITY
INT) LOCATION 'kafka://localhost:2181/brokers?topic=orders'
CREATE EXTERNAL TABLE LARGE_ORDERS (ID INT PRIMARY KEY, TOTAL INT)
LOCATION 'kafka://localhost:2181/brokers?topic=large_orders' TBLPROPERTIES
'{"producer":{"bootstrap.servers":"localhost:9092","acks":"1","key.serializer":"org.apache.org.apache.storm.kafka.IntSerializer","value.serializer":"org.apache.org.apache.storm.kafka.ByteBufferSerializer"}}'
INSERT INTO LARGE_ORDERS SELECT ID, UNIT_PRICE * QUANTITY AS TOTAL FROM
ORDERS WHERE UNIT_PRICE * QUANTITY > 50
```
-The first statement defines the table `ORDER` which represents the input
stream. The `LOCATION` clause specifies the ZkHost (`localhost:2181`), the path
of the brokers in ZooKeeper (`/brokers`) and the topic (`orders`). The
`TBLPROPERTIES` clause specifies the configuration of
[KafkaProducer](http://kafka.apache.org/documentation.html#producerconfigs).
-Current implementation of `storm-sql-kafka` requires specifying both
`LOCATION` and `TBLPROPERTIES` clauses even though the table is read-only or
write-only.
+The first statement defines the table `ORDER` which represents the input
stream. The `LOCATION` clause specifies the ZkHost (`localhost:2181`), the path
of the brokers in ZooKeeper (`/brokers`) and the topic (`orders`).
-Similarly, the second statement specifies the table `LARGE_ORDERS` which
represents the output stream. The third statement is a `SELECT` statement which
defines the topology: it instructs StormSQL to filter all orders in the
external table `ORDERS`, calculates the total price and inserts matching
records into the Kafka stream specified by `LARGE_ORDER`.
+Similarly, the second statement specifies the table `LARGE_ORDERS` which
represents the output stream. The `TBLPROPERTIES` clause specifies the
configuration of
[KafkaProducer](http://kafka.apache.org/documentation.html#producerconfigs).
--- End diff --
Might be better to clarify again that `TBLPROPERTIES` is optional for table
as a producer (read) and mandatory for table as a consumer (write).
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---