Github user gvramana commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2366#discussion_r195980220
--- Diff:
integration/spark-common/src/main/java/org/apache/carbondata/streaming/CarbonStreamRecordReader.java
---
@@ -71,15 +72,10 @@
import org.apache.hadoop.mapreduce.TaskAttemptContext;
import org.apache.hadoop.mapreduce.lib.input.FileSplit;
import org.apache.spark.memory.MemoryMode;
+import org.apache.carbondata.spark.vectorreader.CarbonSparkVectorReader;
import org.apache.spark.sql.catalyst.InternalRow;
import org.apache.spark.sql.catalyst.expressions.GenericInternalRow;
-import org.apache.spark.sql.execution.vectorized.ColumnVector;
-import org.apache.spark.sql.execution.vectorized.ColumnarBatch;
-import org.apache.spark.sql.types.CalendarIntervalType;
-import org.apache.spark.sql.types.Decimal;
-import org.apache.spark.sql.types.DecimalType;
-import org.apache.spark.sql.types.StructField;
-import org.apache.spark.sql.types.StructType;
+import org.apache.spark.sql.types.*;
--- End diff --
Add another sub-issue and PR for CarbonStreaming spark and hadoop
dependency refactoring.
1) Move CarbonStreamRecordReader.java, Spark2 and
CarbonStreamInputFormat.java to Carbon-hadoop and use CarbonStreamRecordReader
using reflection.
2) Take out dependency of Spark on Carbon-streaming.
---