Dear all,

I'm in a decisive moment where I should have answers to few questions. So any (prompt, even incomplete) answer is highly appreciated.

- What is the overhead of updating the schema? in my case, only adding new columns; knowing that the schema could evolve from few tens to few hundreds of columns. - When scaling to hundreds of columns, do the performances decrease noticeably? - Do the performance of reading from/writing to Parquet files depend on the processing system used like: Impala, Hive, Spark, etc.?

Regards, Grüße, Cordialement, Recuerdos, Saluti, προσρήσεις, 问候, تحياتي.
Mohamed Nadjib Mami
PhD Student - EIS Group- Bonn University (Germany).
About me! <http://www.strikingly.com/mohamed-nadjib-mami> LinkedIn <http://fr.linkedin.com/in/mohamednadjibmami/>

Reply via email to