Test Result

40 failures (+38) , 66 skipped (-618)
18,236 tests (-4393)
Took 13 hr.

All Failed Tests

Test NameDurationAge
 org.apache.spark.scheduler.BarrierTaskContextSuite.throw exception on barrier() call timeout20 sec1
 org.apache.spark.sql.kafka010.KafkaContinuousSinkSuite.streaming - write w/o topic field, with topic option1 min 8 sec1
 org.apache.spark.sql.kafka010.KafkaContinuousSinkSuite.generic - write big data with small producer buffer7 min 59 sec1
 org.apache.spark.sql.kafka010.KafkaDontFailOnDataLossSuite.failOnDataLoss=false should not return duplicated records: microbatch v143 sec1
 org.apache.spark.sql.kafka010.KafkaDontFailOnDataLossSuite.failOnDataLoss=false should not return duplicated records: microbatch v234 sec1
 org.apache.spark.sql.kafka010.KafkaDontFailOnDataLossSuite.failOnDataLoss=false should not return duplicated records: continuous processing55 sec1
 org.apache.spark.sql.kafka010.KafkaDontFailOnDataLossSuite.failOnDataLoss=false should not return duplicated records: batch v11 min 55 sec1
 org.apache.spark.sql.kafka010.KafkaDontFailOnDataLossSuite.failOnDataLoss=false should not return duplicated records: batch v245 sec1
 org.apache.spark.sql.kafka010.KafkaMicroBatchV1SourceSuite.subscribing topic by pattern from earliest offsets (failOnDataLoss: false)1 min 26 sec1
 org.apache.spark.sql.kafka010.KafkaMicroBatchV1SourceSuite.subscribing topic by pattern from specific offsets (failOnDataLoss: false)1 min 39 sec1
 org.apache.spark.sql.kafka010.KafkaMicroBatchV2SourceSuite.subscribing topic by pattern with topic deletions1 min 24 sec1
 org.apache.spark.sql.kafka010.KafkaMicroBatchV2SourceSuite.subscribe topic by pattern with topic recreation between batches1 min 0 sec1
 org.apache.spark.sql.kafka010.KafkaMicroBatchV2SourceSuite.read Kafka transactional messages: read_uncommitted1 min 0 sec1
 org.apache.spark.sql.kafka010.KafkaMicroBatchV2SourceSuite.reset should reset all fields1 min 0 sec1
 org.apache.spark.sql.kafka010.KafkaMicroBatchV2SourceSuite.413 sec1
 org.apache.spark.sql.kafka010.KafkaOffsetReaderSuite.SPARK-30656: getOffsetRangesFromUnresolvedOffsets - using specific offsets28 sec1
 org.apache.spark.sql.kafka010.KafkaOffsetReaderSuite.SPARK-30656: getOffsetRangesFromUnresolvedOffsets - using special offsets1 min 0 sec1
 org.apache.spark.sql.kafka010.KafkaOffsetReaderSuite.SPARK-30656: getOffsetRangesFromUnresolvedOffsets - multiple topic partitions1 min 0 sec1
 org.apache.spark.sql.kafka010.KafkaOffsetReaderSuite.SPARK-30656: getOffsetRangesFromResolvedOffsets1 min 0 sec1
 org.apache.spark.sql.kafka010.KafkaSinkBatchSuiteV1.batch - write to kafka1 min 5 sec1
 org.apache.spark.sql.kafka010.KafkaSinkBatchSuiteV1.batch - partition column and partitioner priorities1 min 4 sec1
 org.apache.spark.sql.kafka010.KafkaSinkBatchSuiteV1.SPARK-20496: batch - enforce analyzed plans2 min 38 sec1
 org.apache.spark.sql.kafka010.KafkaSinkBatchSuiteV1.batch - unsupported save modes1 min 4 sec1
 org.apache.spark.sql.kafka010.KafkaSinkBatchSuiteV2.(It is not a test it is a sbt.testing.SuiteSelector)54 sec1
 org.apache.spark.sql.kafka010.KafkaSourceStressForDontFailOnDataLossSuite.stress test for failOnDataLoss=false1 min 6 sec1
 org.apache.spark.sql.kafka010.consumer.KafkaDataConsumerSuite.SPARK-25151 Handles multiple tasks in executor fetching same (topic, partition) pair59 sec1
 org.apache.spark.sql.kafka010.consumer.KafkaDataConsumerSuite.SPARK-25151 Handles multiple tasks in executor fetching same (topic, partition) pair and same offset (edge-case) - data in use4 min 0 sec1
 org.apache.spark.streaming.kafka010.DirectKafkaStreamSuite.offset recovery38 sec1
 org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.(It is not a test it is a sbt.testing.SuiteSelector)13 min1
 org.apache.spark.sql.hive.HiveSparkSubmitSuite.SPARK-18989: DESC TABLE should not fail with format class not found5 min 0 sec1
 org.apache.spark.sql.hive.client.VersionsSuite.0: createDatabase48 sec1
 org.apache.spark.sql.hive.client.VersionsSuite.0: create/get/alter database should pick right user name as owner6 ms1
 org.apache.spark.sql.hive.client.VersionsSuite.0: createDatabase with null description28 ms1
 org.apache.spark.sql.hive.client.VersionsSuite.0: alterDatabase5 ms1
 org.apache.spark.sql.hive.client.VersionsSuite.0: dropDatabase0.24 sec1
 org.apache.spark.sql.hive.client.VersionsSuite.0: alterTable - change database7 ms1
 org.apache.spark.sql.hive.client.VersionsSuite.0: alterTable - change database and table names5 ms1
 org.apache.spark.sql.hive.client.VersionsSuite.0: listTables(database)0.14 sec1
 org.apache.spark.sql.hive.client.VersionsSuite.0: listTables(database, pattern)17 ms1
 org.apache.spark.sql.hive.client.VersionsSuite.0: dropTable1.7 sec1

All Tests

PackageDurationFail(diff)Skip(diff)Pass(diff)Total(diff)
org.apache.spark35 min01454455
org.apache.spark.api.java0.19 sec0099
org.apache.spark.api.python3.4 sec0088
org.apache.spark.api.r24 ms0033
org.apache.spark.broadcast1 min 6 sec001717
org.apache.spark.deploy4 min 3 sec019495
org.apache.spark.deploy.client0.77 sec0022
org.apache.spark.deploy.history2 min 6 sec00165165
org.apache.spark.deploy.k8s0.16 sec001717
org.apache.spark.deploy.k8s.features4.8 sec005353
org.apache.spark.deploy.k8s.submit0.64 sec001111
org.apache.spark.deploy.master53 sec003030
org.apache.spark.deploy.master.ui0.26 sec0022
org.apache.spark.deploy.mesos0.2 sec0055
org.apache.spark.deploy.rest4.3 sec002727
org.apache.spark.deploy.rest.mesos5 sec0033
org.apache.spark.deploy.security46 sec0055
org.apache.spark.deploy.worker7.4 sec003232
org.apache.spark.deploy.worker.ui0.22 sec0011
org.apache.spark.deploy.yarn27 sec009898
org.apache.spark.executor3.5 sec003232
org.apache.spark.graphx1 min 29 sec005757
org.apache.spark.graphx.impl0.27 sec001818
org.apache.spark.graphx.lib7 min 49 sec002727
org.apache.spark.graphx.util6.3 sec0066
org.apache.spark.input0.46 sec0033
org.apache.spark.internal16 ms0011
org.apache.spark.internal.config0.21 sec002727
org.apache.spark.internal.io7 ms0066
org.apache.spark.internal.plugin35 sec0055
org.apache.spark.io5.6 sec004040
org.apache.spark.kafka0104 sec004545
org.apache.spark.launcher1 min 21 sec005353
org.apache.spark.memory0.85 sec002929
org.apache.spark.metrics2.8 sec003535
org.apache.spark.metrics.sink0.15 sec0066
org.apache.spark.metrics.source0.84 sec0077
org.apache.spark.ml18 sec0016-316-3
org.apache.spark.ml.classification9 min 12 sec00-285-11185-113
org.apache.spark.ml.clustering1 min 16 sec009-519-51
org.apache.spark.ml.evaluation4.8 sec005-225-22
org.apache.spark.ml.feature1 min 40 sec0095-28395-283
org.apache.spark.ml.fpm59 sec007-47-4
org.apache.spark.ml.impl22 ms0011
org.apache.spark.ml.linalg1.3 sec0085-685-6
org.apache.spark.ml.optim1.1 sec003-93-9
org.apache.spark.ml.param.shared0.12 sec0011
org.apache.spark.ml.recommendation12 sec002-342-34
org.apache.spark.ml.regression5 min 19 sec02-167-5669-57
org.apache.spark.ml.source.libsvm4.1 sec0011-111-1
org.apache.spark.ml.stat.distribution0.2 sec0044
org.apache.spark.ml.util4 sec0011-1111-11
org.apache.spark.mllib.classification40 sec005-415-41
org.apache.spark.mllib.clustering29 sec0013-5113-51
org.apache.spark.mllib.evaluation1.7 sec0010-1210-12
org.apache.spark.mllib.feature27 sec006-306-30
org.apache.spark.mllib.fpm8 sec008-138-13
org.apache.spark.mllib.linalg0.71 sec009-879-87
org.apache.spark.mllib.pmml.export74 ms003-83-8
org.apache.spark.mllib.random6.3 sec007-177-17
org.apache.spark.mllib.recommendation2.4 sec005-215-21
org.apache.spark.mllib.stat2.7 sec016-287-28
org.apache.spark.mllib.tree36 sec006-336-33
org.apache.spark.mllib.util3 ms002-232-23
org.apache.spark.network42 sec004242
org.apache.spark.network.crypto2.7 sec001717
org.apache.spark.network.netty1.8 sec001414
org.apache.spark.network.protocol47 ms0066
org.apache.spark.network.sasl3.5 sec001616
org.apache.spark.network.server73 ms0033
org.apache.spark.network.shuffle6.5 sec006161
org.apache.spark.network.util21 sec001010
org.apache.spark.network.yarn25 sec001313
org.apache.spark.partial0.15 sec001010
org.apache.spark.rdd2 min 33 sec00241241
org.apache.spark.repl4 min 13 sec004040
org.apache.spark.resource39 sec002727
org.apache.spark.rpc13 ms0055
org.apache.spark.rpc.netty12 sec005050
org.apache.spark.scheduler9 min 46 sec1+10317-1318
org.apache.spark.scheduler.cluster7.5 sec0022
org.apache.spark.scheduler.cluster.k8s6 sec002323
org.apache.spark.scheduler.cluster.mesos47 sec009999
org.apache.spark.scheduler.dynalloc0.19 sec001212
org.apache.spark.security20 sec001010
org.apache.spark.serializer30 sec00110110
org.apache.spark.shuffle28 sec0055
org.apache.spark.shuffle.sort4.6 sec004848
org.apache.spark.shuffle.sort.io13 ms0022
org.apache.spark.sql2 hr 22 min0627682774
org.apache.spark.sql.api.python8.8 sec0022
org.apache.spark.sql.api.r0.25 sec0011
org.apache.spark.sql.avro3 min 27 sec00194194
org.apache.spark.sql.catalyst1.9 sec004949
org.apache.spark.sql.catalyst.analysis31 sec00555555
org.apache.spark.sql.catalyst.catalog15 sec00160160
org.apache.spark.sql.catalyst.csv1.6 sec003333
org.apache.spark.sql.catalyst.encoders47 sec00380380
org.apache.spark.sql.catalyst.expressions23 min00804804
org.apache.spark.sql.catalyst.expressions.aggregate19 sec005050
org.apache.spark.sql.catalyst.expressions.codegen1 min 9 sec006767
org.apache.spark.sql.catalyst.expressions.xml8.5 sec002020
org.apache.spark.sql.catalyst.json0.17 sec001515
org.apache.spark.sql.catalyst.optimizer24 sec01615616
org.apache.spark.sql.catalyst.parser8.5 sec00263263
org.apache.spark.sql.catalyst.planning27 ms0088
org.apache.spark.sql.catalyst.plans0.87 sec003838
org.apache.spark.sql.catalyst.plans.logical29 ms001313
org.apache.spark.sql.catalyst.statsEstimation1.3 sec00114114
org.apache.spark.sql.catalyst.streaming7 ms0022
org.apache.spark.sql.catalyst.trees1.4 sec002929
org.apache.spark.sql.catalyst.util10 sec02241243
org.apache.spark.sql.connector1 min 54 sec00419-97419-97
org.apache.spark.sql.connector.catalog3.3 sec007373
org.apache.spark.sql.connector.expressions55 ms0066
org.apache.spark.sql.execution18 min04543-3547-3
org.apache.spark.sql.execution.adaptive38 sec003131
org.apache.spark.sql.execution.aggregate0.92 sec0011
org.apache.spark.sql.execution.arrow7.1 sec003535
org.apache.spark.sql.execution.columnar15 sec00155-32155-32
org.apache.spark.sql.execution.columnar.compression3.9 sec00118118
org.apache.spark.sql.execution.command5 min 35 sec00242242
org.apache.spark.sql.execution.datasources4 min 10 sec00226226
org.apache.spark.sql.execution.datasources.binaryfile4.2 sec0099
org.apache.spark.sql.execution.datasources.csv6 min 14 sec00354354
org.apache.spark.sql.execution.datasources.jdbc5 ms0011
org.apache.spark.sql.execution.datasources.json2 min 33 sec02336338
org.apache.spark.sql.execution.datasources.noop1.3 sec0066
org.apache.spark.sql.execution.datasources.orc10 min01313314
org.apache.spark.sql.execution.datasources.parquet42 min01-4481-76482-80
org.apache.spark.sql.execution.datasources.text2 min 23 sec004040
org.apache.spark.sql.execution.datasources.v23.3 sec006161
org.apache.spark.sql.execution.debug1.3 sec0077
org.apache.spark.sql.execution.history0.59 sec0022
org.apache.spark.sql.execution.joins4 min 33 sec02133-48135-48
org.apache.spark.sql.execution.metric50 sec002727
org.apache.spark.sql.execution.python18 sec002525
org.apache.spark.sql.execution.streaming46 sec015455
org.apache.spark.sql.execution.streaming.continuous.shuffle89 ms001616
org.apache.spark.sql.execution.streaming.sources44 sec004242
org.apache.spark.sql.execution.streaming.state11 min004242
org.apache.spark.sql.execution.ui16 sec001414
org.apache.spark.sql.execution.vectorized57 sec006262
org.apache.spark.sql.expressions7.3 sec0088
org.apache.spark.sql.hive1 hr 31 min2+20-1388-234390-233
org.apache.spark.sql.hive.client1 hr 28 min10+101856-46867-36
org.apache.spark.sql.hive.execution1 hr 1 min01-589409-1126410-1715
org.apache.spark.sql.hive.orc4 min 8 sec01-136-10937-110
org.apache.spark.sql.internal1 min 44 sec007575
org.apache.spark.sql.jdbc36 sec01115116
org.apache.spark.sql.kafka0101 hr 41 min24+240213-29237-5
org.apache.spark.sql.kafka010.consumer15 min2+2016-218
org.apache.spark.sql.kafka010.producer1.3 sec0044
org.apache.spark.sql.sources12 min02-2247-137249-139
org.apache.spark.sql.streaming16 min00385-24385-24
org.apache.spark.sql.streaming.continuous2 min 17 sec004141
org.apache.spark.sql.streaming.sources12 sec002727
org.apache.spark.sql.streaming.test37 sec002626
org.apache.spark.sql.streaming.ui32 sec0077
org.apache.spark.sql.test1 min 30 sec006060
org.apache.spark.sql.types0.36 sec02191193
org.apache.spark.sql.util9.7 sec014445
org.apache.spark.status9.7 sec003131
org.apache.spark.status.api.v13 ms0011
org.apache.spark.storage37 sec00238238
org.apache.spark.streaming9 min 49 sec01243244
org.apache.spark.streaming.api.java56 ms0011
org.apache.spark.streaming.kafka0102 min 50 sec1+1024-125
org.apache.spark.streaming.kinesis2.7 sec0243761
org.apache.spark.streaming.rdd18 sec001313
org.apache.spark.streaming.receiver0.54 sec0066
org.apache.spark.streaming.scheduler27 sec002323
org.apache.spark.streaming.scheduler.rate2.4 sec001111
org.apache.spark.streaming.ui3.6 sec0088
org.apache.spark.streaming.util18 sec005555
org.apache.spark.ui52 sec024547
org.apache.spark.ui.scope3 ms0011
org.apache.spark.ui.storage71 ms0055
org.apache.spark.unsafe59 ms0088
org.apache.spark.unsafe.array1 ms0011
org.apache.spark.unsafe.hash0.54 sec0066
org.apache.spark.unsafe.map9.4 sec003030
org.apache.spark.unsafe.types1.4 sec016162
org.apache.spark.util1 min 8 sec00174174
org.apache.spark.util.collection6 min 41 sec03123126
org.apache.spark.util.collection.unsafe.sort9.5 sec008282
org.apache.spark.util.io40 ms0088
org.apache.spark.util.kvstore8.7 sec01104105
org.apache.spark.util.logging0.85 sec0011
org.apache.spark.util.random7.1 sec003232
org.apache.spark.util.sketch18 sec002929
test.org.apache.spark1 min 53 sec00109109
test.org.apache.spark.streaming1 min 13 sec007979