Console Output

Skipping 23,904 KB.. Full Log
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:77)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:137)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:58)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431)
[info]   2020-05-15 11:12:13.697 - stderr> 	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
[info]   2020-05-15 11:12:13.697 - stderr> 	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
[info]   2020-05-15 11:12:13.697 - stderr> 	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[info]   2020-05-15 11:12:13.697 - stderr> 	at java.base/java.lang.reflect.Method.invoke(Method.java:566)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:79)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:164)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:70)
[info]   2020-05-15 11:12:13.697 - stderr> 	at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
[info]   2020-05-15 11:12:13.697 - stderr> 	at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
[info]   2020-05-15 11:12:13.697 - stderr> 	at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
[info]   2020-05-15 11:12:13.697 - stderr> 	at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:83)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248)
[info]   2020-05-15 11:12:13.697 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:388)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221)
[info]   2020-05-15 11:12:13.698 - stderr> 	at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.init(SparkSQLEnv.scala:58)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.<init>(SparkSQLCLIDriver.scala:321)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:155)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
[info]   2020-05-15 11:12:13.698 - stderr> 	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
[info]   2020-05-15 11:12:13.698 - stderr> 	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
[info]   2020-05-15 11:12:13.698 - stderr> 	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[info]   2020-05-15 11:12:13.698 - stderr> 	at java.base/java.lang.reflect.Method.invoke(Method.java:566)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:934)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1013)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1022)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
[info]   2020-05-15 11:12:13.698 - stderr> Caused by: ERROR XJ041: Failed to create database '/home/jenkins/workspace/NewSparkPullRequestBuilder/target/tmp/spark-e622abb8-8c37-49c0-a881-da9238419759', see the next exception for details.
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
[info]   2020-05-15 11:12:13.698 - stderr> 	... 109 more
[info]   2020-05-15 11:12:13.698 - stderr> Caused by: ERROR XBM0J: Directory /home/jenkins/workspace/NewSparkPullRequestBuilder/target/tmp/spark-e622abb8-8c37-49c0-a881-da9238419759 already exists.
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
[info]   2020-05-15 11:12:13.698 - stderr> 	at java.base/java.security.AccessController.doPrivileged(Native Method)
[info]   2020-05-15 11:12:13.698 - stderr> 	at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.jdbc.EmbedConnection$5.run(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at java.base/java.security.AccessController.doPrivileged(Native Method)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.jdbc.EmbedConnection.createPersistentService(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	... 106 more
[info]   2020-05-15 11:12:13.699 - stderr> ------
[info]   2020-05-15 11:12:13.699 - stderr> 
[info]   2020-05-15 11:12:13.699 - stderr> 	at jdk.internal.reflect.GeneratedConstructorAccessor69.newInstance(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
[info] - FMClassifier doesn't fit linear when fitLinear is off (1 second, 995 milliseconds)
[info]   2020-05-15 11:12:13.699 - stderr> 	at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)
[info]   2020-05-15 11:12:13.699 - stderr> 	at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192)
[info]   2020-05-15 11:12:13.699 - stderr> 	at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:422)
[info]   2020-05-15 11:12:13.699 - stderr> 	at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:297)
[info]   2020-05-15 11:12:13.699 - stderr> 	at jdk.internal.reflect.GeneratedConstructorAccessor67.newInstance(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
[info]   2020-05-15 11:12:13.699 - stderr> 	at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817)
[info]   2020-05-15 11:12:13.699 - stderr> 	... 82 more
[info]   2020-05-15 11:12:13.699 - stderr> Caused by: java.sql.SQLException: Failed to create database '/home/jenkins/workspace/NewSparkPullRequestBuilder/target/tmp/spark-e622abb8-8c37-49c0-a881-da9238419759', see the next exception for details.
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at java.base/java.security.AccessController.doPrivileged(Native Method)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
[info]   2020-05-15 11:12:13.699 - stderr> 	at java.sql/java.sql.DriverManager.getConnection(DriverManager.java:677)
[info]   2020-05-15 11:12:13.699 - stderr> 	at java.sql/java.sql.DriverManager.getConnection(DriverManager.java:189)
[info]   2020-05-15 11:12:13.699 - stderr> 	at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
[info]   2020-05-15 11:12:13.699 - stderr> 	at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
[info]   2020-05-15 11:12:13.699 - stderr> 	... 93 more
[info]   2020-05-15 11:12:13.699 - stderr> Caused by: ERROR XJ041: Failed to create database '/home/jenkins/workspace/NewSparkPullRequestBuilder/target/tmp/spark-e622abb8-8c37-49c0-a881-da9238419759', see the next exception for details.
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	... 109 more
[info]   2020-05-15 11:12:13.7 - stderr> Caused by: ERROR XBM0J: Directory /home/jenkins/workspace/NewSparkPullRequestBuilder/target/tmp/spark-e622abb8-8c37-49c0-a881-da9238419759 already exists.
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at java.base/java.security.AccessController.doPrivileged(Native Method)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.impl.jdbc.EmbedConnection$5.run(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	at java.base/java.security.AccessController.doPrivileged(Native Method)
[info]   2020-05-15 11:12:13.7 - stderr> 	at org.apache.derby.impl.jdbc.EmbedConnection.createPersistentService(Unknown Source)
[info]   2020-05-15 11:12:13.7 - stderr> 	... 106 more
[info]   ===========================
[info]   End CliSuite failure output
[info]   =========================== (CliSuite.scala:192)
[info]   org.scalatest.exceptions.TestFailedException:
[info]   at org.scalatest.Assertions.newAssertionFailedException(Assertions.scala:530)
[info]   at org.scalatest.Assertions.newAssertionFailedException$(Assertions.scala:529)
[info]   at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1560)
[info]   at org.scalatest.Assertions.fail(Assertions.scala:1107)
[info]   at org.scalatest.Assertions.fail$(Assertions.scala:1103)
[info]   at org.scalatest.FunSuite.fail(FunSuite.scala:1560)
[info]   at org.apache.spark.sql.hive.thriftserver.CliSuite.runCliWithin(CliSuite.scala:192)
[info]   at org.apache.spark.sql.hive.thriftserver.CliSuite.$anonfun$new$19(CliSuite.scala:436)
[info]   at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
[info]   at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
[info]   at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
[info]   at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
[info]   at org.scalatest.Transformer.apply(Transformer.scala:22)
[info]   at org.scalatest.Transformer.apply(Transformer.scala:20)
[info]   at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186)
[info]   at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:151)
[info]   at org.scalatest.FunSuiteLike.invokeWithFixture$1(FunSuiteLike.scala:184)
[info]   at org.scalatest.FunSuiteLike.$anonfun$runTest$1(FunSuiteLike.scala:196)
[info]   at org.scalatest.SuperEngine.runTestImpl(Engine.scala:286)
[info]   at org.scalatest.FunSuiteLike.runTest(FunSuiteLike.scala:196)
[info]   at org.scalatest.FunSuiteLike.runTest$(FunSuiteLike.scala:178)
[info]   at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterEach$$super$runTest(SparkFunSuite.scala:58)
[info]   at org.scalatest.BeforeAndAfterEach.runTest(BeforeAndAfterEach.scala:221)
[info]   at org.scalatest.BeforeAndAfterEach.runTest$(BeforeAndAfterEach.scala:214)
[info]   at org.apache.spark.SparkFunSuite.runTest(SparkFunSuite.scala:58)
[info]   at org.scalatest.FunSuiteLike.$anonfun$runTests$1(FunSuiteLike.scala:229)
[info]   at org.scalatest.SuperEngine.$anonfun$runTestsInBranch$1(Engine.scala:393)
[info]   at scala.collection.immutable.List.foreach(List.scala:392)
[info]   at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:381)
[info]   at org.scalatest.SuperEngine.runTestsInBranch(Engine.scala:376)
[info]   at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:458)
[info]   at org.scalatest.FunSuiteLike.runTests(FunSuiteLike.scala:229)
[info]   at org.scalatest.FunSuiteLike.runTests$(FunSuiteLike.scala:228)
[info]   at org.scalatest.FunSuite.runTests(FunSuite.scala:1560)
[info]   at org.scalatest.Suite.run(Suite.scala:1124)
[info]   at org.scalatest.Suite.run$(Suite.scala:1106)
[info]   at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1560)
[info]   at org.scalatest.FunSuiteLike.$anonfun$run$1(FunSuiteLike.scala:233)
[info]   at org.scalatest.SuperEngine.runImpl(Engine.scala:518)
[info]   at org.scalatest.FunSuiteLike.run(FunSuiteLike.scala:233)
[info]   at org.scalatest.FunSuiteLike.run$(FunSuiteLike.scala:232)
[info]   at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:58)
[info]   at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:213)
[info]   at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
[info]   at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
[info]   at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:58)
[info]   at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:317)
[info]   at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:510)
[info]   at sbt.ForkMain$Run$2.call(ForkMain.java:296)
[info]   at sbt.ForkMain$Run$2.call(ForkMain.java:286)
[info]   at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
[info]   at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
[info]   at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
[info]   at java.base/java.lang.Thread.run(Thread.java:834)
[info]   Cause: java.util.concurrent.TimeoutException: Futures timed out after [1 minute]
[info]   at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:259)
[info]   at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:263)
[info]   at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:294)
[info]   at org.apache.spark.sql.hive.thriftserver.CliSuite.runCliWithin(CliSuite.scala:174)
[info]   at org.apache.spark.sql.hive.thriftserver.CliSuite.$anonfun$new$19(CliSuite.scala:436)
[info]   at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
[info]   at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
[info]   at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
[info]   at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
[info]   at org.scalatest.Transformer.apply(Transformer.scala:22)
[info]   at org.scalatest.Transformer.apply(Transformer.scala:20)
[info]   at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186)
[info]   at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:151)
[info]   at org.scalatest.FunSuiteLike.invokeWithFixture$1(FunSuiteLike.scala:184)
[info]   at org.scalatest.FunSuiteLike.$anonfun$runTest$1(FunSuiteLike.scala:196)
[info]   at org.scalatest.SuperEngine.runTestImpl(Engine.scala:286)
[info]   at org.scalatest.FunSuiteLike.runTest(FunSuiteLike.scala:196)
[info]   at org.scalatest.FunSuiteLike.runTest$(FunSuiteLike.scala:178)
[info]   at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterEach$$super$runTest(SparkFunSuite.scala:58)
[info]   at org.scalatest.BeforeAndAfterEach.runTest(BeforeAndAfterEach.scala:221)
[info]   at org.scalatest.BeforeAndAfterEach.runTest$(BeforeAndAfterEach.scala:214)
[info]   at org.apache.spark.SparkFunSuite.runTest(SparkFunSuite.scala:58)
[info]   at org.scalatest.FunSuiteLike.$anonfun$runTests$1(FunSuiteLike.scala:229)
[info]   at org.scalatest.SuperEngine.$anonfun$runTestsInBranch$1(Engine.scala:393)
[info]   at scala.collection.immutable.List.foreach(List.scala:392)
[info]   at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:381)
[info]   at org.scalatest.SuperEngine.runTestsInBranch(Engine.scala:376)
[info]   at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:458)
[info]   at org.scalatest.FunSuiteLike.runTests(FunSuiteLike.scala:229)
[info]   at org.scalatest.FunSuiteLike.runTests$(FunSuiteLike.scala:228)
[info]   at org.scalatest.FunSuite.runTests(FunSuite.scala:1560)
[info]   at org.scalatest.Suite.run(Suite.scala:1124)
[info]   at org.scalatest.Suite.run$(Suite.scala:1106)
[info]   at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1560)
[info]   at org.scalatest.FunSuiteLike.$anonfun$run$1(FunSuiteLike.scala:233)
[info]   at org.scalatest.SuperEngine.runImpl(Engine.scala:518)
[info]   at org.scalatest.FunSuiteLike.run(FunSuiteLike.scala:233)
[info]   at org.scalatest.FunSuiteLike.run$(FunSuiteLike.scala:232)
[info]   at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:58)
[info]   at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:213)
[info]   at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
[info]   at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
[info]   at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:58)
[info]   at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:317)
[info]   at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:510)
[info]   at sbt.ForkMain$Run$2.call(ForkMain.java:296)
[info]   at sbt.ForkMain$Run$2.call(ForkMain.java:286)
[info]   at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
[info]   at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
[info]   at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
[info]   at java.base/java.lang.Thread.run(Thread.java:834)
[info] - prediction on single instance (2 seconds, 74 milliseconds)
[info] - read/write (2 seconds, 131 milliseconds)
[info] LeastSquaresAggregatorSuite:
[info] - aggregator add method input size (26 milliseconds)
[info] - negative weight (25 milliseconds)
[info] - check sizes (56 milliseconds)
[info] - check correctness (393 milliseconds)
[info] - check with zero standard deviation (56 milliseconds)
[info] LabeledPointSuite:
[info] - parse labeled points (2 milliseconds)
[info] - parse labeled points with whitespaces (1 millisecond)
[info] - parse labeled points with v0.9 format (1 millisecond)
[info] - conversions between new ml LabeledPoint and mllib LabeledPoint (1 millisecond)
[info] - Kryo class register (96 milliseconds)
[info] NaiveBayesSuite:
[info] - model types (43 milliseconds)
[info] - get, set params (29 milliseconds)
[info] - Naive Bayes Multinomial (1 second, 384 milliseconds)
[info] - Naive Bayes Bernoulli (1 second, 45 milliseconds)
[info] - detect negative values (363 milliseconds)
[info] - detect non zero or one values in Bernoulli (770 milliseconds)
[info] - model save/load: 2.0 to 2.0 (1 second, 483 milliseconds)
[info] - model save/load: 1.0 to 2.0 (1 second, 590 milliseconds)
[info] NaiveBayesClusterSuite:
[info] - write bucketed data with sortBy (31 seconds, 93 milliseconds)
[info] - write bucketed data with the overlapping bucketBy/sortBy and partitionBy columns (7 milliseconds)
[info] - write bucketed data without partitionBy (16 seconds, 58 milliseconds)
[info] - write bucketed data without partitionBy with sortBy (40 seconds, 583 milliseconds)
Build timed out (after 300 minutes). Marking the build as failed.
Build was aborted
Archiving artifacts
Recording test results
Finished: FAILURE