[jira] [Commented] (SPARK-19288) Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in R/run-tests.sh
[ https://issues.apache.org/jira/browse/SPARK-19288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15931586#comment-15931586 ] Felix Cheung commented on SPARK-19288: -- If he's on IST, it's exactly 12hr 30min ahead of PST (where Jenkins is running), which could explain the offset of 30min. I think we need to fix up these tests to account of time zone differences > Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in > R/run-tests.sh > -- > > Key: SPARK-19288 > URL: https://issues.apache.org/jira/browse/SPARK-19288 > Project: Spark > Issue Type: Bug > Components: SparkR, SQL, Tests >Affects Versions: 2.0.1 > Environment: Ubuntu 16.04, X86_64, ppc64le >Reporter: Nirman Narang > > Full log here. > {code:title=R/run-tests.sh|borderStyle=solid} > Loading required package: methods > Attaching package: 'SparkR' > The following object is masked from 'package:testthat': > describe > The following objects are masked from 'package:stats': > cov, filter, lag, na.omit, predict, sd, var, window > The following objects are masked from 'package:base': > as.data.frame, colnames, colnames<-, drop, intersect, rank, rbind, > sample, subset, summary, transform, union > functions on binary files : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > > binary functions : ... > broadcast variables : .. > functions in client.R : . > test functions in sparkR.R : .Re-using existing Spark Context. Call > sparkR.session.stop() or restart R to create a new Spark Context > ... > include R packages : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > JVM API : .. > MLlib functions : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > .SLF4J: Failed to load class > "org.slf4j.impl.StaticLoggerBinder". > SLF4J: Defaulting to no-operation (NOP) logger implementation > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further > details. > .Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 65,622 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 70B for [label] > BINARY: 1 values, 21B raw, 23B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, > RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 87B for [terms, > list, element, list, element] BINARY: 2 values, 42B raw, 43B comp, 1 pages, > encodings: [PLAIN, RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 30B for > [hasIntercept] BOOLEAN: 1 values, 1B raw, 3B comp, 1 pages, encodings: > [PLAIN, BIT_PACKED] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 49 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 90B for [labels, > list, element] BINARY: 3 values, 50B raw, 50B comp, 1 pages, encodings: > [PLAIN, RLE] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec
[jira] [Commented] (SPARK-19288) Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in R/run-tests.sh
[ https://issues.apache.org/jira/browse/SPARK-19288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15929438#comment-15929438 ] Hyukjin Kwon commented on SPARK-19288: -- FWIW, for me it has been fine. Mac OS 10.12.3 & KTS & R version 3.2.3. It has been fine for Windows via AppVeyor so far. > Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in > R/run-tests.sh > -- > > Key: SPARK-19288 > URL: https://issues.apache.org/jira/browse/SPARK-19288 > Project: Spark > Issue Type: Bug > Components: SparkR, SQL, Tests >Affects Versions: 2.0.1 > Environment: Ubuntu 16.04, X86_64, ppc64le >Reporter: Nirman Narang > > Full log here. > {code:title=R/run-tests.sh|borderStyle=solid} > Loading required package: methods > Attaching package: 'SparkR' > The following object is masked from 'package:testthat': > describe > The following objects are masked from 'package:stats': > cov, filter, lag, na.omit, predict, sd, var, window > The following objects are masked from 'package:base': > as.data.frame, colnames, colnames<-, drop, intersect, rank, rbind, > sample, subset, summary, transform, union > functions on binary files : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > > binary functions : ... > broadcast variables : .. > functions in client.R : . > test functions in sparkR.R : .Re-using existing Spark Context. Call > sparkR.session.stop() or restart R to create a new Spark Context > ... > include R packages : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > JVM API : .. > MLlib functions : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > .SLF4J: Failed to load class > "org.slf4j.impl.StaticLoggerBinder". > SLF4J: Defaulting to no-operation (NOP) logger implementation > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further > details. > .Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 65,622 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 70B for [label] > BINARY: 1 values, 21B raw, 23B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, > RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 87B for [terms, > list, element, list, element] BINARY: 2 values, 42B raw, 43B comp, 1 pages, > encodings: [PLAIN, RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 30B for > [hasIntercept] BOOLEAN: 1 values, 1B raw, 3B comp, 1 pages, encodings: > [PLAIN, BIT_PACKED] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 49 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 90B for [labels, > list, element] BINARY: 3 values, 50B raw, 50B comp, 1 pages, encodings: > [PLAIN, RLE] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apa
[jira] [Commented] (SPARK-19288) Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in R/run-tests.sh
[ https://issues.apache.org/jira/browse/SPARK-19288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15929433#comment-15929433 ] Miao Wang commented on SPARK-19288: --- I think it only happens at local build. I had another similar issue due to building hive once. So if you comment out this one, it will pass. I don't think Jenkins will suffer this issue. > Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in > R/run-tests.sh > -- > > Key: SPARK-19288 > URL: https://issues.apache.org/jira/browse/SPARK-19288 > Project: Spark > Issue Type: Bug > Components: SparkR, SQL, Tests >Affects Versions: 2.0.1 > Environment: Ubuntu 16.04, X86_64, ppc64le >Reporter: Nirman Narang > > Full log here. > {code:title=R/run-tests.sh|borderStyle=solid} > Loading required package: methods > Attaching package: 'SparkR' > The following object is masked from 'package:testthat': > describe > The following objects are masked from 'package:stats': > cov, filter, lag, na.omit, predict, sd, var, window > The following objects are masked from 'package:base': > as.data.frame, colnames, colnames<-, drop, intersect, rank, rbind, > sample, subset, summary, transform, union > functions on binary files : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > > binary functions : ... > broadcast variables : .. > functions in client.R : . > test functions in sparkR.R : .Re-using existing Spark Context. Call > sparkR.session.stop() or restart R to create a new Spark Context > ... > include R packages : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > JVM API : .. > MLlib functions : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > .SLF4J: Failed to load class > "org.slf4j.impl.StaticLoggerBinder". > SLF4J: Defaulting to no-operation (NOP) logger implementation > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further > details. > .Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 65,622 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 70B for [label] > BINARY: 1 values, 21B raw, 23B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, > RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 87B for [terms, > list, element, list, element] BINARY: 2 values, 42B raw, 43B comp, 1 pages, > encodings: [PLAIN, RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 30B for > [hasIntercept] BOOLEAN: 1 values, 1B raw, 3B comp, 1 pages, encodings: > [PLAIN, BIT_PACKED] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 49 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 90B for [labels, > list, element] BINARY: 3 values, 50B raw, 50B comp, 1 pages, encodings: > [PLAIN, RLE] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig:
[jira] [Commented] (SPARK-19288) Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in R/run-tests.sh
[ https://issues.apache.org/jira/browse/SPARK-19288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15833608#comment-15833608 ] Nirman Narang commented on SPARK-19288: --- [~felixcheung] I am using R version 3.2.3. Tested on X86_64 and ppc64le with Ubuntu 16.04 and it fails on both. > Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in > R/run-tests.sh > -- > > Key: SPARK-19288 > URL: https://issues.apache.org/jira/browse/SPARK-19288 > Project: Spark > Issue Type: Bug > Components: SparkR, SQL, Tests >Affects Versions: 2.0.1 > Environment: Ubuntu 16.04, X86_64, ppc64le >Reporter: Nirman Narang > > Full log here. > {code:title=R/run-tests.sh|borderStyle=solid} > Loading required package: methods > Attaching package: 'SparkR' > The following object is masked from 'package:testthat': > describe > The following objects are masked from 'package:stats': > cov, filter, lag, na.omit, predict, sd, var, window > The following objects are masked from 'package:base': > as.data.frame, colnames, colnames<-, drop, intersect, rank, rbind, > sample, subset, summary, transform, union > functions on binary files : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > > binary functions : ... > broadcast variables : .. > functions in client.R : . > test functions in sparkR.R : .Re-using existing Spark Context. Call > sparkR.session.stop() or restart R to create a new Spark Context > ... > include R packages : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > JVM API : .. > MLlib functions : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > .SLF4J: Failed to load class > "org.slf4j.impl.StaticLoggerBinder". > SLF4J: Defaulting to no-operation (NOP) logger implementation > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further > details. > .Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 65,622 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 70B for [label] > BINARY: 1 values, 21B raw, 23B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, > RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 87B for [terms, > list, element, list, element] BINARY: 2 values, 42B raw, 43B comp, 1 pages, > encodings: [PLAIN, RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 30B for > [hasIntercept] BOOLEAN: 1 values, 1B raw, 3B comp, 1 pages, encodings: > [PLAIN, BIT_PACKED] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 49 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 90B for [labels, > list, element] BINARY: 3 values, 50B raw, 50B comp, 1 pages, encodings: > [PLAIN, RLE] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.par
[jira] [Commented] (SPARK-19288) Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in R/run-tests.sh
[ https://issues.apache.org/jira/browse/SPARK-19288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15832837#comment-15832837 ] Felix Cheung commented on SPARK-19288: -- hmm, that's odd. what system and R version? I'm wondering if this is related to time zone? > Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in > R/run-tests.sh > -- > > Key: SPARK-19288 > URL: https://issues.apache.org/jira/browse/SPARK-19288 > Project: Spark > Issue Type: Bug > Components: SparkR, SQL, Tests >Affects Versions: 2.0.1 > Environment: Ubuntu 16.04, X86_64, ppc64le >Reporter: Nirman Narang > > Full log here. > {code:title=R/run-tests.sh|borderStyle=solid} > Loading required package: methods > Attaching package: 'SparkR' > The following object is masked from 'package:testthat': > describe > The following objects are masked from 'package:stats': > cov, filter, lag, na.omit, predict, sd, var, window > The following objects are masked from 'package:base': > as.data.frame, colnames, colnames<-, drop, intersect, rank, rbind, > sample, subset, summary, transform, union > functions on binary files : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > > binary functions : ... > broadcast variables : .. > functions in client.R : . > test functions in sparkR.R : .Re-using existing Spark Context. Call > sparkR.session.stop() or restart R to create a new Spark Context > ... > include R packages : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > JVM API : .. > MLlib functions : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > .SLF4J: Failed to load class > "org.slf4j.impl.StaticLoggerBinder". > SLF4J: Defaulting to no-operation (NOP) logger implementation > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further > details. > .Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 65,622 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 70B for [label] > BINARY: 1 values, 21B raw, 23B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, > RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 87B for [terms, > list, element, list, element] BINARY: 2 values, 42B raw, 43B comp, 1 pages, > encodings: [PLAIN, RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 30B for > [hasIntercept] BOOLEAN: 1 values, 1B raw, 3B comp, 1 pages, encodings: > [PLAIN, BIT_PACKED] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 49 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 90B for [labels, > list, element] BINARY: 3 values, 50B raw, 50B comp, 1 pages, encodings: > [PLAIN, RLE] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutp
[jira] [Commented] (SPARK-19288) Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in R/run-tests.sh
[ https://issues.apache.org/jira/browse/SPARK-19288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15831748#comment-15831748 ] Nirman Narang commented on SPARK-19288: --- [~felixcheung] Version 2.0.1 > Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in > R/run-tests.sh > -- > > Key: SPARK-19288 > URL: https://issues.apache.org/jira/browse/SPARK-19288 > Project: Spark > Issue Type: Bug > Components: SparkR, SQL, Tests >Affects Versions: 2.0.1 > Environment: Ubuntu 16.04, X86_64, ppc64le >Reporter: Nirman Narang > > Full log here. > {code:title=R/run-tests.sh|borderStyle=solid} > Loading required package: methods > Attaching package: 'SparkR' > The following object is masked from 'package:testthat': > describe > The following objects are masked from 'package:stats': > cov, filter, lag, na.omit, predict, sd, var, window > The following objects are masked from 'package:base': > as.data.frame, colnames, colnames<-, drop, intersect, rank, rbind, > sample, subset, summary, transform, union > functions on binary files : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > > binary functions : ... > broadcast variables : .. > functions in client.R : . > test functions in sparkR.R : .Re-using existing Spark Context. Call > sparkR.session.stop() or restart R to create a new Spark Context > ... > include R packages : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > JVM API : .. > MLlib functions : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > .SLF4J: Failed to load class > "org.slf4j.impl.StaticLoggerBinder". > SLF4J: Defaulting to no-operation (NOP) logger implementation > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further > details. > .Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 65,622 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 70B for [label] > BINARY: 1 values, 21B raw, 23B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, > RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 87B for [terms, > list, element, list, element] BINARY: 2 values, 42B raw, 43B comp, 1 pages, > encodings: [PLAIN, RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 30B for > [hasIntercept] BOOLEAN: 1 values, 1B raw, 3B comp, 1 pages, encodings: > [PLAIN, BIT_PACKED] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 49 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 90B for [labels, > list, element] BINARY: 3 values, 50B raw, 50B comp, 1 pages, encodings: > [PLAIN, RLE] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017
[jira] [Commented] (SPARK-19288) Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in R/run-tests.sh
[ https://issues.apache.org/jira/browse/SPARK-19288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15830284#comment-15830284 ] Felix Cheung commented on SPARK-19288: -- We are not seeing this in Jenkins? Which branch are you running this from? > Failure (at test_sparkSQL.R#1300): date functions on a DataFrame in > R/run-tests.sh > -- > > Key: SPARK-19288 > URL: https://issues.apache.org/jira/browse/SPARK-19288 > Project: Spark > Issue Type: Bug > Components: SparkR, SQL, Tests >Affects Versions: 2.0.1 > Environment: Ubuntu 16.04, X86_64, ppc64le >Reporter: Nirman Narang > > Full log here. > {code:title=R/run-tests.sh|borderStyle=solid} > Loading required package: methods > Attaching package: 'SparkR' > The following object is masked from 'package:testthat': > describe > The following objects are masked from 'package:stats': > cov, filter, lag, na.omit, predict, sd, var, window > The following objects are masked from 'package:base': > as.data.frame, colnames, colnames<-, drop, intersect, rank, rbind, > sample, subset, summary, transform, union > functions on binary files : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > > binary functions : ... > broadcast variables : .. > functions in client.R : . > test functions in sparkR.R : .Re-using existing Spark Context. Call > sparkR.session.stop() or restart R to create a new Spark Context > ... > include R packages : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > JVM API : .. > MLlib functions : Spark package found in SPARK_HOME: > /var/lib/jenkins/workspace/Sparkv2.0.1/spark > .SLF4J: Failed to load class > "org.slf4j.impl.StaticLoggerBinder". > SLF4J: Defaulting to no-operation (NOP) logger implementation > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further > details. > .Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:53 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 65,622 > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 70B for [label] > BINARY: 1 values, 21B raw, 23B comp, 1 pages, encodings: [PLAIN, BIT_PACKED, > RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 87B for [terms, > list, element, list, element] BINARY: 2 values, 42B raw, 43B comp, 1 pages, > encodings: [PLAIN, RLE] > Jan 19, 2017 5:40:54 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 30B for > [hasIntercept] BOOLEAN: 1 values, 1B raw, 3B comp, 1 pages, encodings: > [PLAIN, BIT_PACKED] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet block size to 134217728 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Parquet dictionary page size to 1048576 > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Dictionary is on > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Validation is off > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Writer version is: PARQUET_1_0 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.InternalParquetRecordWriter: Flushing mem > columnStore to file. allocated memory: 49 > Jan 19, 2017 5:40:55 PM INFO: > org.apache.parquet.hadoop.ColumnChunkPageWriteStore: written 90B for [labels, > list, element] BINARY: 3 values, 50B raw, 50B comp, 1 pages, encodings: > [PLAIN, RLE] > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.codec.CodecConfig: > Compression: SNAPPY > Jan 19, 2017 5:40:55 PM INFO: org.apache.parquet.hadoop.ParquetOutputFormat: > Par