[GitHub] [hudi] pratyakshsharma commented on pull request #2424: [HUDI-1509]: Reverting LinkedHashSet changes to fix performance degradation for large schemas

2021-01-11 Thread GitBox
pratyakshsharma commented on pull request #2424: URL: https://github.com/apache/hudi/pull/2424#issuecomment-758461542 @n3nash In my previous org, we were dealing with a similar scenario where fields were getting deleted from few tables in production. Yeah parquet-avro reader will throw

[GitHub] [hudi] pratyakshsharma commented on pull request #2424: [HUDI-1509]: Reverting LinkedHashSet changes to fix performance degradation for large schemas

2021-01-11 Thread GitBox
pratyakshsharma commented on pull request #2424: URL: https://github.com/apache/hudi/pull/2424#issuecomment-758162231 @n3nash Just a high level thought before going through the changes thoroughly. How about keeping the old changes also and introduce a config

[GitHub] [hudi] pratyakshsharma commented on pull request #2424: [HUDI-1509]: Reverting LinkedHashSet changes to fix performance degradation for large schemas

2021-01-11 Thread GitBox
pratyakshsharma commented on pull request #2424: URL: https://github.com/apache/hudi/pull/2424#issuecomment-758154222 > @n3nash what is the commit being reverted? https://github.com/apache/hudi/commit/6d7ca2cf7e441ad19d32d7a25739e454f39ed253