Re: Path style access fs.s3a.path.style.access property is not working in spark code

2020-05-04 Thread Aniruddha P Tekade
Hello User, I got the solution to this. If you are writing to a custom s3 url, then use the hadoop-aws-2.8.0.jar as the separate flag was introduced to enable path style access. Best, Aniruddha --- ᐧ On Fri, May 1, 2020 at 5:08 PM Aniruddha P Tekade wrote: > Hello Users, >

Path style access fs.s3a.path.style.access property is not working in spark code

2020-05-01 Thread Aniruddha P Tekade
Hello Users, I am using on-premise object storage and able to perform operations on different bucket using aws-cli. However, when I am trying to use the same path from my spark code, it fails. Here are the details - Addes dependencies in build.sbt - - hadoop-aws-2.7.4.ja -

Spark job stuck at s3a-file-system metrics system started

2020-04-29 Thread Aniruddha P Tekade
Hello, I am trying to run a spark job that is trying to write the data into a custom s3 endpoint bucket. But I am stuck at this line of output and job is not moving forward at all - 20/04/29 16:03:59 INFO SharedState: Setting hive.metastore.warehouse.dir ('null') to the value of

Re: [External Email] Re: Standard practices for building dashboards for spark processed data

2020-02-26 Thread Aniruddha P Tekade
for the client to post > process. > > Kind regards, > > Aniruddha P Tekade schrieb am Mi. 26. Feb. 2020 > um 02:23: > >> Hello, >> >> I am trying to build a data pipeline that uses spark structured streaming >> with delta project and runs into Kub

Standard practices for building dashboards for spark processed data

2020-02-25 Thread Aniruddha P Tekade
Hello, I am trying to build a data pipeline that uses spark structured streaming with delta project and runs into Kubernetes. Due to this, I get my output files only into parquet format. Since I am asked to use the prometheus and grafana for building the dashboard for this pipeline, I run an

spark writeStream not working with custom S3 endpoint

2019-12-03 Thread Aniruddha P Tekade
Hello, While working with Spark Structured Streaming (v2.4.3) I am trying to write my streaming dataframe to a custom S3. I have made sure that I am able to login, upload data to s3 buckets manually using UI and have also setup ACCESS_KEY and SECRET_KEY for it. val sc = spark.sparkContext

Can spark convert String to Integer when reading using schema in structured streaming

2019-11-22 Thread Aniruddha P Tekade
Hi, I am new to spark and learning spark structured streaming. I am using structured streaming with schema specified with the help of case class and encoders to get the streaming dataframe. case class SampleLogEntry( dateTime: Timestamp,