This is an automated email from the ASF dual-hosted git repository.
bridgetb pushed a commit to branch gh-pages
in repository https://gitbox.apache.org/repos/asf/drill.git
The following commit(s) were added to refs/heads/gh-pages by this push:
new c1cf52f Update 2014-12-09-running-sql-queries-on-amazon-s3.md (#164)
c1cf52f is described below
commit c1cf52f386690cfc85eed30851efcef6911c61db
Author: Steven Phillips <[email protected]>
AuthorDate: Fri Jun 1 13:16:26 2018 -0700
Update 2014-12-09-running-sql-queries-on-amazon-s3.md (#164)
---
blog/_posts/2014-12-09-running-sql-queries-on-amazon-s3.md | 8 +++++---
1 file changed, 5 insertions(+), 3 deletions(-)
diff --git a/blog/_posts/2014-12-09-running-sql-queries-on-amazon-s3.md
b/blog/_posts/2014-12-09-running-sql-queries-on-amazon-s3.md
index a8a5781..d1edd1b 100644
--- a/blog/_posts/2014-12-09-running-sql-queries-on-amazon-s3.md
+++ b/blog/_posts/2014-12-09-running-sql-queries-on-amazon-s3.md
@@ -70,10 +70,12 @@ Next, let's go into the Drill web interface and enable the
S3 storage plugin. I
1. Point your browser to `http://<host>:8047`
2. Select the 'Storage' tab.
-2. A good starting configuration for S3 can be entirely the same as the `dfs`
plugin, except the connection parameter is changed to `s3://bucket`. So first
select the `Update` button for `dfs`, then select the text area and copy it
into the clipboard (on Windows, ctrl-A, ctrl-C works).
+2. A good starting configuration for S3 can be entirely the same as the `dfs`
plugin, except the connection parameter is changed to `s3n://bucket`. So first
select the `Update` button for `dfs`, then select the text area and copy it
into the clipboard (on Windows, ctrl-A, ctrl-C works).
2. Press `Back`, then create a new plugin by typing the name into the `New
Storage Plugin`, then press `Create`. You can choose any name, but a good
convention is to use `s3-<bucketname>` so you can easily identify it later.
-3. In the configuration area, paste the configuration you just grabbed from
'dfs'. Change the line `connection: "file:///"` to `connection:
"s3://<bucket>"`.
+3. In the configuration area, paste the configuration you just grabbed from
'dfs'. Change the line `connection: "file:///"` to `connection:
"s3n://<bucket>"`.
4. Click `Update`. You should see a message that indicates success.
+
+Note: Make sure the URI has scheme "s3n", not "s3". It will not work with "s3".
At this point you can run queries on the data directly and you have a couple
of options on how you want to access it. You can use Drill Explorer and create
a custom view (based on an SQL query) that you can then access in Tableau or
other BI tools, or just use Drill directly from within the tool.
@@ -81,4 +83,4 @@ You may want to check out the [Tableau
demo](http://www.youtube.com/watch?v=jNUs
With just a few lines of configuration, you've just opened the vast world of
data available in the Amazon cloud and reduced the amount of work you have to
do in advance to access data stored there with SQL. There are even some
[public datasets](https://aws.amazon.com/datasets) available directly on S3
that are great for experimentation.
-Happy Drilling!
\ No newline at end of file
+Happy Drilling!
--
To stop receiving notification emails like this one, please contact
[email protected].