Dear all,
I'm in a decisive moment where I should have answers to few questions.
So any (prompt, even incomplete) answer is highly appreciated.
- What is the overhead of updating the schema? in my case, only adding
new columns; knowing that the schema could evolve from few tens to few
hundreds of columns.
- When scaling to hundreds of columns, do the performances decrease
noticeably?
- Do the performance of reading from/writing to Parquet files depend on
the processing system used like: Impala, Hive, Spark, etc.?
Regards, Grüße, Cordialement, Recuerdos, Saluti, προσρήσεις, 问候, تحياتي.
Mohamed Nadjib Mami
PhD Student - EIS Group- Bonn University (Germany).
About me! <http://www.strikingly.com/mohamed-nadjib-mami> LinkedIn
<http://fr.linkedin.com/in/mohamednadjibmami/>