gunjdesai opened a new issue #4437:
URL: https://github.com/apache/hudi/issues/4437


   I am using `Spark Structured Streaming (3.1.1)` to read data from `Kafka` 
and use `HUDI (0.8.0)` as the storage system on S3 partitioning the data by 
date. (no problems with this section)
   
   I am looking to use `Trino (355)` to be able to query that data. As a 
pre-curser, I've already placed the `hudi-presto-bundle-0.8.0.jar` in 
`/data/trino/hive/`
   
   I created a table with the following schema
   
   ```
   CREATE TABLE table_new (
     columns, dt
   ) WITH (
     partitioned_by = ARRAY['dt'], 
     external_location = 's3a://bucket/location/',
     format = 'parquet'
   );
   ```
   Even after calling the below function, trino is unable to discover any 
partitions
   
   ```
   CALL system.sync_partition_metadata('schema', 'table_new', 'ALL')
   ```
   My assessment is that I am unable to create a table under trino using hudi 
largely due to the fact that I am not able to pass the right values under 
`WITH` Options. I am also unable to find a create table example under 
documentation for HUDI.
   
   I would really appreciate if anyone can give me a example for that, or point 
me to the right direction, if in case I've missed anything.
   
   Really appreciate the help
   
   **Environment Description**
   
   * Hudi version : 0.8.0
   
   * Spark version : 3.1.1
   
   * Hive version : 3.0.0 (Metastore only)
   
   * Hadoop version : 3.2.0
   
   * Storage (HDFS/S3/GCS..) : S3
   
   * Running on Docker? (yes/no) : Yes
   
   This is more of a question than an issue, sorry couldn't apply a label for 
it.
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to