FailedConsole Output

Skipping 15,346 KB.. Full Log
ationWriterThread$$anonfun$run$1.apply$mcV$sp(ScriptTransformationExec.scala:281)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1988)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.run(ScriptTransformationExec.scala:270)
16:39:31.744 ERROR org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread: 
16:39:31.745 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 15513.0 (TID 36333, localhost, executor driver): java.lang.IllegalArgumentException: intentional exception
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:146)
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:143)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
Exception in thread "Thread-ScriptTransformation-Feed" 	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
java.lang.IllegalArgumentException: intentional exception
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:146)
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:143)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply$mcV$sp(ScriptTransformationExec.scala:281)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1988)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.run(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply$mcV$sp(ScriptTransformationExec.scala:281)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1988)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.run(ScriptTransformationExec.scala:270)

16:39:31.745 ERROR org.apache.spark.scheduler.TaskSetManager: Task 0 in stage 15513.0 failed 1 times; aborting job
16:39:31.745 ERROR org.apache.spark.util.Utils: Uncaught exception in thread Thread-ScriptTransformation-Feed
java.lang.IllegalArgumentException: intentional exception
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:146)
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:143)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply$mcV$sp(ScriptTransformationExec.scala:281)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1988)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.run(ScriptTransformationExec.scala:270)
[info] - script transformation should not swallow errors from upstream operators (no serde) (1 second, 165 milliseconds)
16:39:32.918 ERROR org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread: 
16:39:32.920 ERROR org.apache.spark.executor.Executor: Exception in task 0.0 in stage 15514.0 (TID 36334)
Exception in thread "Thread-ScriptTransformation-Feed" java.lang.IllegalArgumentException: intentional exception
java.lang.IllegalArgumentException: intentional exception
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:146)
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:143)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply$mcV$sp(ScriptTransformationExec.scala:281)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1988)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.run(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:146)
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:143)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply$mcV$sp(ScriptTransformationExec.scala:281)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1988)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.run(ScriptTransformationExec.scala:270)
16:39:32.920 ERROR org.apache.spark.util.Utils: Uncaught exception in thread Thread-ScriptTransformation-Feed
java.lang.IllegalArgumentException: intentional exception
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:146)
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:143)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply$mcV$sp(ScriptTransformationExec.scala:281)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1988)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.run(ScriptTransformationExec.scala:270)
16:39:32.924 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 15514.0 (TID 36334, localhost, executor driver): java.lang.IllegalArgumentException: intentional exception
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:146)
	at org.apache.spark.sql.hive.execution.ExceptionInjectingOperator$$anonfun$doExecute$1.apply(ScriptTransformationSuite.scala:143)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply$mcV$sp(ScriptTransformationExec.scala:281)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread$$anonfun$run$1.apply(ScriptTransformationExec.scala:270)
	at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1988)
	at org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.run(ScriptTransformationExec.scala:270)

16:39:32.925 ERROR org.apache.spark.scheduler.TaskSetManager: Task 0 in stage 15514.0 failed 1 times; aborting job
[info] - script transformation should not swallow errors from upstream operators (with serde) (1 second, 170 milliseconds)
16:39:33.083 ERROR org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread: /bin/bash: some_non_existent_command: command not found

16:39:33.086 ERROR org.apache.spark.sql.hive.execution.ScriptTransformationExec: /bin/bash: some_non_existent_command: command not found

16:39:33.087 ERROR org.apache.spark.sql.hive.execution.ScriptTransformationExec: /bin/bash: some_non_existent_command: command not found

16:39:33.087 ERROR org.apache.spark.executor.Executor: Exception in task 0.0 in stage 15515.0 (TID 36335)
org.apache.spark.SparkException: Subprocess exited with status 127. Error: /bin/bash: some_non_existent_command: command not found

	at org.apache.spark.sql.hive.execution.ScriptTransformationExec$$anon$1.checkFailureAndPropagate(ScriptTransformationExec.scala:144)
	at org.apache.spark.sql.hive.execution.ScriptTransformationExec$$anon$1.hasNext(ScriptTransformationExec.scala:189)
	at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
	at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:239)
	at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:233)
	at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:828)
	at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:828)
	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:324)
	at org.apache.spark.rdd.RDD.iterator(RDD.scala:288)
	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
	at org.apache.spark.scheduler.Task.run(Task.scala:109)
	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.spark.SparkException: Subprocess exited with status 127. Error: /bin/bash: some_non_existent_command: command not found

	at org.apache.spark.sql.hive.execution.ScriptTransformationExec$$anon$1.checkFailureAndPropagate(ScriptTransformationExec.scala:144)
	at org.apache.spark.sql.hive.execution.ScriptTransformationExec$$anon$1.hasNext(ScriptTransformationExec.scala:178)
	... 14 more
16:39:33.090 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 15515.0 (TID 36335, localhost, executor driver): org.apache.spark.SparkException: Subprocess exited with status 127. Error: /bin/bash: some_non_existent_command: command not found

	at org.apache.spark.sql.hive.execution.ScriptTransformationExec$$anon$1.checkFailureAndPropagate(ScriptTransformationExec.scala:144)
	at org.apache.spark.sql.hive.execution.ScriptTransformationExec$$anon$1.hasNext(ScriptTransformationExec.scala:189)
	at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
	at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:239)
	at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:233)
	at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:828)
	at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:828)
	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:324)
	at org.apache.spark.rdd.RDD.iterator(RDD.scala:288)
	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
	at org.apache.spark.scheduler.Task.run(Task.scala:109)
	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.spark.SparkException: Subprocess exited with status 127. Error: /bin/bash: some_non_existent_command: command not found

	at org.apache.spark.sql.hive.execution.ScriptTransformationExec$$anon$1.checkFailureAndPropagate(ScriptTransformationExec.scala:144)
	at org.apache.spark.sql.hive.execution.ScriptTransformationExec$$anon$1.hasNext(ScriptTransformationExec.scala:178)
	... 14 more

16:39:33.090 ERROR org.apache.spark.scheduler.TaskSetManager: Task 0 in stage 15515.0 failed 1 times; aborting job
[info] - SPARK-14400 script transformation should fail for bad script command (164 milliseconds)
[info] BigDataBenchmarkSuite:
[info] - No data files found for BigDataBenchmark tests. !!! IGNORED !!!
[info] ExpressionSQLBuilderSuite:
[info] - literal (2 milliseconds)
[info] - attributes (0 milliseconds)
[info] - binary comparisons (1 millisecond)
[info] - logical operators (0 milliseconds)
[info] - arithmetic expressions (1 millisecond)
[info] - window specification (0 milliseconds)
[info] - interval arithmetic (3 milliseconds)
[info] QueryPartitionSuite:
[info] - SPARK-5068: query data when path doesn't exist (3 seconds, 765 milliseconds)
16:39:36.993 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/table_with_timestamp_partition specified for non-external table:table_with_timestamp_partition
[info] - SPARK-21739: Cast expression should initialize timezoneId (1 second, 241 milliseconds)
[info] ConcurrentHiveSuite:
[info] - multiple instances not supported !!! IGNORED !!!
[info] ShowCreateTableSuite:
16:39:38.319 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider json. Persisting data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
16:39:38.515 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider json. Persisting data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
[info] - data source table with user specified schema (340 milliseconds)
16:39:38.907 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider json. Persisting data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
16:39:38.961 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider json. Persisting data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
[info] - data source table CTAS (448 milliseconds)
16:39:39.426 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider json. Persisting data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
16:39:39.564 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider json. Persisting data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
[info] - partitioned data source table (600 milliseconds)
16:39:40.029 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Persisting bucketed data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive. 
16:39:40.086 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Persisting bucketed data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive. 
[info] - bucketed data source table (513 milliseconds)
16:39:40.540 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Persisting bucketed data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive. 
16:39:40.664 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Persisting bucketed data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive. 
[info] - partitioned bucketed data source table (591 milliseconds)
16:39:41.341 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Persisting bucketed data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive. 
16:39:41.502 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Persisting bucketed data source table `default`.`ddl_test` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive. 
[info] - data source table using Dataset API (827 milliseconds)
16:39:41.638 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
16:39:41.791 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
[info] - simple hive table (289 milliseconds)
[info] - simple external hive table (92 milliseconds)
16:39:42.019 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
16:39:42.168 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
[info] - partitioned hive table (285 milliseconds)
16:39:42.305 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
16:39:42.453 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
[info] - hive table with explicit storage info (288 milliseconds)
16:39:42.593 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
16:39:42.746 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
[info] - hive table with STORED AS clause (292 milliseconds)
16:39:42.886 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
16:39:43.024 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
[info] - hive table with serde info (283 milliseconds)
[info] - hive view (88 milliseconds)
[info] - hive view with output columns (80 milliseconds)
16:39:43.339 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
16:39:43.493 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
[info] - hive bucketing is supported (307 milliseconds)
16:39:43.647 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/t1 specified for non-external table:t1
[info] - hive partitioned view is not supported (272 milliseconds)
[info] HiveTableScanSuite:
16:39:44.051 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/src specified for non-external table:src
16:39:44.398 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/part_scan_test specified for non-external table:part_scan_test
[info] - partition_based_table_scan_with_different_serde (2 seconds, 147 milliseconds)
16:39:46.317 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/src specified for non-external table:src
[info] - file_split_for_small_table (693 milliseconds)
16:39:46.843 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/tb specified for non-external table:tb
[info] - Spark-4041: lowercase issue (503 milliseconds)
16:39:47.346 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/timestamp_query_null specified for non-external table:timestamp_query_null
[info] - Spark-4077: timestamp query for null value (495 milliseconds)
16:39:47.840 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/spark_4959 specified for non-external table:spark_4959
[info] - Spark-4959 Attributes are case sensitive when using a select query from a projection (950 milliseconds)
16:39:48.850 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/table_with_partition specified for non-external table:table_with_partition
[info] - Verify SQLConf HIVE_METASTORE_PARTITION_PRUNING (1 second, 810 milliseconds)
16:39:50.608 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/table_with_partition specified for non-external table:table_with_partition
[info] - SPARK-16926: number of table and partition columns match for new partitioned table (1 second, 572 milliseconds)
16:39:52.180 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/hive_tbl_part specified for non-external table:hive_tbl_part
[info] - HiveTableScanExec canonicalization for different orders of partition filters (179 milliseconds)
[info] ListTablesSuite:
16:39:52.678 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/hivelisttablessuitetable specified for non-external table:hivelisttablessuitetable
16:39:52.748 WARN org.apache.hadoop.hive.metastore.ObjectStore: Failed to get database listtablessuitedb, returning NoSuchObjectException
16:39:52.836 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/listtablessuitedb.db/hiveindblisttablessuitetable specified for non-external table:hiveindblisttablessuitetable
[info] - get all tables of current database (400 milliseconds)
[info] - getting all tables with a database name (279 milliseconds)
[info] HiveExplainSuite:
16:39:53.725 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/src specified for non-external table:src
[info] - show cost in explain command (529 milliseconds)
[info] - explain extended command (59 milliseconds)
[info] - explain create table command (53 milliseconds)
[info] - SPARK-17409: The EXPLAIN output of CTAS only shows the analyzed plan (85 milliseconds)
[info] - EXPLAIN CODEGEN command (31 milliseconds)
[info] UDFSuite:
16:39:54.636 WARN org.apache.hadoop.hive.metastore.HiveMetaStore: Location: file:/home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/target/tmp/warehouse-f2d27724-6b3c-4126-8c37-3c1eead3a832/src specified for non-external table:src
[info] - UDF case insensitive (517 milliseconds)
[info] - temporary function: create and drop (53 milliseconds)
[info] - permanent function: create and drop without specifying db name (110 milliseconds)
[info] - permanent function: create and drop with a db name (124 milliseconds)
16:39:55.442 WARN org.apache.hadoop.hive.metastore.ObjectStore: Failed to get database db_638fc450_31dd_4a45_823f_0a4a39c6b7f6, returning NoSuchObjectException
[info] - permanent function: create and drop a function in another db (274 milliseconds)
[info] Test run started
[info] Test org.apache.spark.sql.hive.JavaMetastoreDataSourcesSuite.saveExternalTableAndQueryIt started
16:39:56.289 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider org.apache.spark.sql.json. Persisting data source table `default`.`javasavedtable` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
16:39:56.540 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider org.apache.spark.sql.json. Persisting data source table `default`.`externaltable` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
[info] Test org.apache.spark.sql.hive.JavaMetastoreDataSourcesSuite.saveTableAndQueryIt started
16:39:57.048 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider org.apache.spark.sql.json. Persisting data source table `default`.`javasavedtable` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
[info] Test org.apache.spark.sql.hive.JavaMetastoreDataSourcesSuite.saveExternalTableWithSchemaAndQueryIt started
16:39:57.414 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider org.apache.spark.sql.json. Persisting data source table `default`.`javasavedtable` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
16:39:57.576 WARN org.apache.spark.sql.hive.test.TestHiveExternalCatalog: Couldn't find corresponding Hive SerDe for data source provider org.apache.spark.sql.json. Persisting data source table `default`.`externaltable` into Hive metastore in Spark SQL specific format, which is NOT compatible with Hive.
[info] Test run finished: 0 failed, 0 ignored, 3 total, 2.151s
[info] Test run started
[info] Test org.apache.spark.sql.hive.JavaDataFrameSuite.testUDAF started
[info] Test org.apache.spark.sql.hive.JavaDataFrameSuite.saveTableAndQueryIt started
[info] Test run finished: 0 failed, 0 ignored, 2 total, 2.29s
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 14 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 8 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 12 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 15 seconds.
[info] Total number of tests run: 5
[info] Suites: completed 1, aborted 0
[info] Tests: succeeded 5, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 5, Failed 0, Errors 0, Passed 5
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 15 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] Passed: Total 101, Failed 0, Errors 0, Passed 100, Skipped 1
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 15 seconds.
[info] Total number of tests run: 29
[info] Suites: completed 3, aborted 0
[info] Tests: succeeded 29, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 29, Failed 0, Errors 0, Passed 29
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 12 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 12 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 12 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 8 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 15 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] Passed: Total 40, Failed 0, Errors 0, Passed 40
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 15 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] Passed: Total 73, Failed 0, Errors 0, Passed 73
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 15 seconds.
[info] Total number of tests run: 19
[info] Suites: completed 1, aborted 0
[info] Tests: succeeded 19, failed 0, canceled 0, ignored 1, pending 0
[info] All tests passed.
[info] Passed: Total 72, Failed 0, Errors 0, Passed 72, Ignored 1
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 15 seconds.
[info] Total number of tests run: 84
[info] Suites: completed 8, aborted 0
[info] Tests: succeeded 84, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 84, Failed 0, Errors 0, Passed 84
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 15 seconds.
[info] Total number of tests run: 0
[info] Suites: completed 0, aborted 0
[info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
[info] No tests were executed.
[info] Passed: Total 41, Failed 0, Errors 0, Passed 41
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 11 seconds.
[info] Total number of tests run: 4
[info] Suites: completed 2, aborted 0
[info] Tests: succeeded 4, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 6, Failed 0, Errors 0, Passed 6
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 11 seconds.
[info] Total number of tests run: 2036
[info] Suites: completed 207, aborted 0
[info] Tests: succeeded 2036, failed 0, canceled 0, ignored 8, pending 0
[info] All tests passed.
[error] Failed: Total 2274, Failed 1, Errors 0, Passed 2273, Ignored 8
[error] Failed tests:
[error] 	test.org.apache.spark.JavaSparkContextSuite
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 11 seconds.
[info] Total number of tests run: 40
[info] Suites: completed 14, aborted 0
[info] Tests: succeeded 40, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 40, Failed 0, Errors 0, Passed 40
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 11 seconds.
[info] Total number of tests run: 83
[info] Suites: completed 9, aborted 0
[info] Tests: succeeded 83, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 83, Failed 0, Errors 0, Passed 83
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 11 seconds.
[info] Total number of tests run: 17
[info] Suites: completed 5, aborted 0
[info] Tests: succeeded 17, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 20, Failed 0, Errors 0, Passed 20
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 11 seconds.
[info] Total number of tests run: 72
[info] Suites: completed 14, aborted 0
[info] Tests: succeeded 72, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 72, Failed 0, Errors 0, Passed 72
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 11 seconds.
[info] Total number of tests run: 14
[info] Suites: completed 2, aborted 0
[info] Tests: succeeded 14, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 18, Failed 0, Errors 0, Passed 18
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 13 seconds.
[info] Total number of tests run: 107
[info] Suites: completed 19, aborted 0
[info] Tests: succeeded 107, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 107, Failed 0, Errors 0, Passed 107
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 13 seconds.
[info] Total number of tests run: 331
[info] Suites: completed 40, aborted 0
[info] Tests: succeeded 331, failed 0, canceled 0, ignored 1, pending 0
[info] All tests passed.
[info] Passed: Total 431, Failed 0, Errors 0, Passed 431, Ignored 1
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 13 seconds.
[info] Total number of tests run: 37
[info] Suites: completed 8, aborted 0
[info] Tests: succeeded 37, failed 0, canceled 0, ignored 24, pending 0
[info] All tests passed.
[info] Passed: Total 44, Failed 0, Errors 0, Passed 44, Ignored 24
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 12 seconds.
[info] Total number of tests run: 2547
[info] Suites: completed 155, aborted 0
[info] Tests: succeeded 2547, failed 0, canceled 0, ignored 2, pending 0
[info] All tests passed.
[info] Passed: Total 2571, Failed 0, Errors 0, Passed 2571, Ignored 2
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 11 seconds.
[info] Total number of tests run: 35
[info] Suites: completed 3, aborted 0
[info] Tests: succeeded 35, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 35, Failed 0, Errors 0, Passed 35
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 10 seconds.
[info] Total number of tests run: 3918
[info] Suites: completed 228, aborted 0
[info] Tests: succeeded 3918, failed 0, canceled 0, ignored 55, pending 0
[info] All tests passed.
[info] Passed: Total 4001, Failed 0, Errors 0, Passed 4001, Ignored 55
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 10 seconds.
[info] Total number of tests run: 66
[info] Suites: completed 9, aborted 0
[info] Tests: succeeded 66, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
[info] Passed: Total 66, Failed 0, Errors 0, Passed 66
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 10 seconds.
[info] Total number of tests run: 1322
[info] Suites: completed 188, aborted 0
[info] Tests: succeeded 1322, failed 0, canceled 0, ignored 7, pending 0
[info] All tests passed.
[info] Passed: Total 1442, Failed 0, Errors 0, Passed 1442, Ignored 7
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 10 seconds.
[info] Total number of tests run: 39
[info] Suites: completed 9, aborted 0
[info] Tests: succeeded 39, failed 0, canceled 0, ignored 2, pending 0
[info] All tests passed.
[info] Passed: Total 39, Failed 0, Errors 0, Passed 39, Ignored 2
[info] ScalaTest
[info] Run completed in 2 hours, 47 minutes, 9 seconds.
[info] Total number of tests run: 2922
[info] Suites: completed 90, aborted 0
[info] Tests: succeeded 2916, failed 6, canceled 0, ignored 595, pending 0
[info] *** 6 TESTS FAILED ***
[error] Failed: Total 2927, Failed 6, Errors 0, Passed 2921, Ignored 595
[error] Failed tests:
[error] 	org.apache.spark.sql.hive.DataSourceWithHiveMetastoreCatalogSuite
[error] (hive/test:test) sbt.TestsFailedException: Tests unsuccessful
[error] (core/test:test) sbt.TestsFailedException: Tests unsuccessful
[error] Total time: 10045 s, completed Jan 3, 2018 4:40:10 PM
[error] running /home/jenkins/workspace/spark-branch-2.3-test-sbt-hadoop-2.6/build/sbt -Phadoop-2.6 -Pkubernetes -Pflume -Phive-thriftserver -Pyarn -Pkafka-0-8 -Phive -Pkinesis-asl -Pmesos test ; received return code 1
Build step 'Execute shell' marked build as failure
Archiving artifacts
Recording test results
Finished: FAILURE