org.apache.spark.AccumulatorSuite: basic accumulation
|
10110 ms |
|
org.apache.spark.AccumulatorSuite: value not assignable from tasks
|
10127 ms |
|
org.apache.spark.AccumulatorSuite: add value to collection accumulators
|
10120 ms |
|
org.apache.spark.AccumulatorSuite: value not readable in tasks
|
10128 ms |
|
org.apache.spark.AccumulatorSuite: collection accumulators
|
10137 ms |
|
org.apache.spark.AccumulatorSuite: localValue readable in tasks
|
10199 ms |
|
org.apache.spark.AccumulatorSuite: garbage collection
|
10125 ms |
|
org.apache.spark.AccumulatorSuite: internal accumulators in TaskContext
|
10289 ms |
|
org.apache.spark.AccumulatorSuite: internal accumulators in a stage
|
10138 ms |
|
org.apache.spark.AccumulatorSuite: internal accumulators in multiple stages
|
10123 ms |
|
org.apache.spark.AccumulatorSuite: internal accumulators in fully resubmitted stages
|
10139 ms |
|
org.apache.spark.AccumulatorSuite: internal accumulators in partially resubmitted stages
|
10127 ms |
|
org.apache.spark.CacheManagerSuite: (It is not a test)
|
10034 ms |
|
org.apache.spark.CheckpointSuite: (It is not a test)
|
11039 ms |
|
org.apache.spark.ContextCleanerSuite: (It is not a test)
|
10622 ms |
|
org.apache.spark.DistributedSuite: task throws not serializable exception
|
10083 ms |
|
org.apache.spark.DistributedSuite: local-cluster format
|
10082 ms |
|
org.apache.spark.DistributedSuite: simple groupByKey
|
10080 ms |
|
org.apache.spark.DistributedSuite: groupByKey where map output sizes exceed maxMbInFlight
|
10134 ms |
|
org.apache.spark.DistributedSuite: accumulators
|
10082 ms |
|
org.apache.spark.DistributedSuite: broadcast variables
|
10085 ms |
|
org.apache.spark.DistributedSuite: repeatedly failing task
|
10121 ms |
|
org.apache.spark.DistributedSuite: repeatedly failing task that crashes JVM
|
10074 ms |
|
org.apache.spark.DistributedSuite: caching
|
10083 ms |
|
org.apache.spark.DistributedSuite: caching on disk
|
10102 ms |
|
org.apache.spark.DistributedSuite: caching in memory, replicated
|
13679 ms |
|
org.apache.spark.DistributedSuite: caching in memory, serialized, replicated
|
10078 ms |
|
org.apache.spark.DistributedSuite: caching on disk, replicated
|
10079 ms |
|
org.apache.spark.DistributedSuite: caching in memory and disk, replicated
|
10087 ms |
|
org.apache.spark.DistributedSuite: caching in memory and disk, serialized, replicated
|
10138 ms |
|
org.apache.spark.DistributedSuite: compute without caching when no partitions fit in memory
|
10363 ms |
|
org.apache.spark.DistributedSuite: compute when only some partitions fit in memory
|
10085 ms |
|
org.apache.spark.DistributedSuite: passing environment variables to cluster
|
10777 ms |
|
org.apache.spark.DistributedSuite: recover from node failures
|
10093 ms |
|
org.apache.spark.DistributedSuite: recover from repeated node failures during shuffle-map
|
10080 ms |
|
org.apache.spark.DistributedSuite: recover from repeated node failures during shuffle-reduce
|
10138 ms |
|
org.apache.spark.DistributedSuite: recover from node failures with replication
|
10114 ms |
|
org.apache.spark.DistributedSuite: unpersist RDDs
|
10087 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: verify min/max executors
|
10146 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: starting state
|
10262 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: add executors
|
10118 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: add executors capped by num pending tasks
|
10118 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: cancel pending executors when no longer needed
|
10150 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: remove executors
|
10133 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: interleaving add and remove
|
10151 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: starting/canceling add timer
|
10271 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: starting/canceling remove timers
|
10143 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: mock polling loop with no events
|
10138 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: mock polling loop add behavior
|
10181 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: mock polling loop remove behavior
|
10202 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: listeners trigger add executors correctly
|
10128 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: listeners trigger remove executors correctly
|
10231 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: listeners trigger add and remove executor callbacks correctly
|
10203 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: SPARK-4951: call onTaskStart before onBlockManagerAdded
|
10869 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: SPARK-4951: onExecutorAdded should not add a busy executor to removeTimes
|
10133 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: avoid ramp up when target < running executors
|
10142 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: avoid ramp down initial executors until first job is submitted
|
10155 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: avoid ramp down initial executors until idle executor is timeout
|
10240 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: get pending task number and related locality preference
|
10135 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: SPARK-8366: maxNumExecutorsNeeded should properly handle failed tasks
|
10745 ms |
|
org.apache.spark.ExecutorAllocationManagerSuite: reset the state of allocation manager
|
10568 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: groupByKey without compression
|
10106 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: shuffle non-zero block size
|
10108 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: shuffle serializer
|
10106 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: zero sized blocks
|
10110 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: zero sized blocks without kryo
|
10189 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: shuffle on mutable pairs
|
10145 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: sorting on mutable pairs
|
10116 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: cogroup using mutable pairs
|
10103 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: subtract mutable pairs
|
10107 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: sort with Java non serializable class - Kryo
|
10093 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: sort with Java non serializable class - Java
|
10112 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: shuffle with different compression settings (SPARK-3426)
|
10099 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: [SPARK-4085] rerun map stage if reduce stage cannot find its local shuffle file
|
10107 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: metrics for shuffle without aggregation
|
10141 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: metrics for shuffle with aggregation
|
10102 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: multiple simultaneous attempts for one task (SPARK-8029)
|
10181 ms |
|
org.apache.spark.ExternalShuffleServiceSuite: using external shuffle service
|
10389 ms |
|
org.apache.spark.FailureSuite: failure in a single-stage job
|
10100 ms |
|
org.apache.spark.FailureSuite: failure in a two-stage job
|
10087 ms |
|
org.apache.spark.FailureSuite: failure in a map stage
|
10150 ms |
|
org.apache.spark.FailureSuite: failure because task results are not serializable
|
10094 ms |
|
org.apache.spark.FailureSuite: failure because task closure is not serializable
|
10088 ms |
|
org.apache.spark.FailureSuite: managed memory leak error should not mask other failures (SPARK-9266
|
10089 ms |
|
org.apache.spark.FailureSuite: last failure cause is sent back to driver
|
10854 ms |
|
org.apache.spark.FailureSuite: failure cause stacktrace is sent back to driver if exception is not serializable
|
10099 ms |
|
org.apache.spark.FailureSuite: failure cause stacktrace is sent back to driver if exception is not deserializable
|
10092 ms |
|
org.apache.spark.FailureSuite: failure in tasks in a submitMapStage
|
10161 ms |
|
org.apache.spark.FileServerSuite: Distributing files locally
|
10035 ms |
|
org.apache.spark.FileServerSuite: Distributing files locally security On
|
10048 ms |
|
org.apache.spark.FileServerSuite: Distributing files locally using URL as input
|
10054 ms |
|
org.apache.spark.FileServerSuite: Dynamically adding JARS locally
|
10062 ms |
|
org.apache.spark.FileServerSuite: Distributing files on a standalone cluster
|
10068 ms |
|
org.apache.spark.FileServerSuite: Dynamically adding JARS on a standalone cluster
|
10298 ms |
|
org.apache.spark.FileServerSuite: Dynamically adding JARS on a standalone cluster using local: URL
|
10042 ms |
|
org.apache.spark.FileSuite: text files
|
10047 ms |
|
org.apache.spark.FileSuite: text files (compressed)
|
10048 ms |
|
org.apache.spark.FileSuite: SequenceFiles
|
10208 ms |
|
org.apache.spark.FileSuite: SequenceFile (compressed)
|
10064 ms |
|
org.apache.spark.FileSuite: SequenceFile with writable key
|
10050 ms |
|
org.apache.spark.FileSuite: SequenceFile with writable value
|
10042 ms |
|
org.apache.spark.FileSuite: SequenceFile with writable key and value
|
10076 ms |
|
org.apache.spark.FileSuite: implicit conversions in reading SequenceFiles
|
10047 ms |
|
org.apache.spark.FileSuite: object files of ints
|
10053 ms |
|
org.apache.spark.FileSuite: object files of complex types
|
10109 ms |
|
org.apache.spark.FileSuite: object files of classes from a JAR
|
10218 ms |
|
org.apache.spark.FileSuite: write SequenceFile using new Hadoop API
|
10128 ms |
|
org.apache.spark.FileSuite: read SequenceFile using new Hadoop API
|
10057 ms |
|
org.apache.spark.FileSuite: binary file input as byte array
|
10065 ms |
|
org.apache.spark.FileSuite: portabledatastream caching tests
|
10046 ms |
|
org.apache.spark.FileSuite: portabledatastream persist disk storage
|
10259 ms |
|
org.apache.spark.FileSuite: portabledatastream flatmap tests
|
10719 ms |
|
org.apache.spark.FileSuite: fixed record length binary file as byte array
|
10502 ms |
|
org.apache.spark.FileSuite: negative binary record length should raise an exception
|
10104 ms |
|
org.apache.spark.FileSuite: file caching
|
10051 ms |
|
org.apache.spark.FileSuite: prevent user from overwriting the empty directory (old Hadoop API)
|
10047 ms |
|
org.apache.spark.FileSuite: prevent user from overwriting the non-empty directory (old Hadoop API)
|
10051 ms |
|
org.apache.spark.FileSuite: allow user to disable the output directory existence checking (old Hadoop API
|
10054 ms |
|
org.apache.spark.FileSuite: prevent user from overwriting the empty directory (new Hadoop API)
|
10052 ms |
|
org.apache.spark.FileSuite: prevent user from overwriting the non-empty directory (new Hadoop API)
|
10051 ms |
|
org.apache.spark.FileSuite: allow user to disable the output directory existence checking (new Hadoop API
|
10054 ms |
|
org.apache.spark.FileSuite: save Hadoop Dataset through old Hadoop API
|
10048 ms |
|
org.apache.spark.FileSuite: save Hadoop Dataset through new Hadoop API
|
10062 ms |
|
org.apache.spark.FileSuite: Get input files via old Hadoop API
|
10086 ms |
|
org.apache.spark.FileSuite: Get input files via new Hadoop API
|
10218 ms |
|
org.apache.spark.FutureActionSuite: (It is not a test)
|
10096 ms |
|
org.apache.spark.HashShuffleSuite: groupByKey without compression
|
10040 ms |
|
org.apache.spark.HashShuffleSuite: shuffle non-zero block size
|
10923 ms |
|
org.apache.spark.HashShuffleSuite: shuffle serializer
|
10038 ms |
|
org.apache.spark.HashShuffleSuite: zero sized blocks
|
13837 ms |
|
org.apache.spark.HashShuffleSuite: zero sized blocks without kryo
|
10466 ms |
|
org.apache.spark.HashShuffleSuite: shuffle on mutable pairs
|
10047 ms |
|
org.apache.spark.HashShuffleSuite: sorting on mutable pairs
|
10408 ms |
|
org.apache.spark.HashShuffleSuite: cogroup using mutable pairs
|
10038 ms |
|
org.apache.spark.HashShuffleSuite: subtract mutable pairs
|
10037 ms |
|
org.apache.spark.HashShuffleSuite: sort with Java non serializable class - Kryo
|
10709 ms |
|
org.apache.spark.HashShuffleSuite: sort with Java non serializable class - Java
|
10037 ms |
|
org.apache.spark.HashShuffleSuite: shuffle with different compression settings (SPARK-3426)
|
10043 ms |
|
org.apache.spark.HashShuffleSuite: [SPARK-4085] rerun map stage if reduce stage cannot find its local shuffle file
|
10049 ms |
|
org.apache.spark.HashShuffleSuite: metrics for shuffle without aggregation
|
10045 ms |
|
org.apache.spark.HashShuffleSuite: metrics for shuffle with aggregation
|
10220 ms |
|
org.apache.spark.HashShuffleSuite: multiple simultaneous attempts for one task (SPARK-8029)
|
10038 ms |
|
org.apache.spark.HeartbeatReceiverSuite: (It is not a test)
|
10078 ms |
|
org.apache.spark.ImplicitOrderingSuite: basic inference of Orderings
|
10455 ms |
|
org.apache.spark.JavaAPISuite: groupByOnPairRDD
|
10164 ms |
|
org.apache.spark.JavaAPISuite: binaryFilesCaching
|
10161 ms |
|
org.apache.spark.JavaAPISuite: sparkContextUnion
|
10237 ms |
|
org.apache.spark.JavaAPISuite: checkpointAndComputation
|
10131 ms |
|
org.apache.spark.JavaAPISuite: leftOuterJoin
|
10138 ms |
|
org.apache.spark.JavaAPISuite: keyByOnPairRDD
|
10893 ms |
|
org.apache.spark.JavaAPISuite: getNumPartitions
|
10147 ms |
|
org.apache.spark.JavaAPISuite: wholeTextFiles
|
10125 ms |
|
org.apache.spark.JavaAPISuite: binaryFiles
|
10166 ms |
|
org.apache.spark.JavaAPISuite: foldReduce
|
10376 ms |
|
org.apache.spark.JavaAPISuite: writeWithNewAPIHadoopFile
|
10164 ms |
|
org.apache.spark.JavaAPISuite: hadoopFile
|
10137 ms |
|
org.apache.spark.JavaAPISuite: lookup
|
10218 ms |
|
org.apache.spark.JavaAPISuite: countAsync
|
10152 ms |
|
org.apache.spark.JavaAPISuite: textFiles
|
10140 ms |
|
org.apache.spark.JavaAPISuite: binaryRecords
|
10227 ms |
|
org.apache.spark.JavaAPISuite: toLocalIterator
|
10734 ms |
|
org.apache.spark.JavaAPISuite: repartitionAndSortWithinPartitions
|
10137 ms |
|
org.apache.spark.JavaAPISuite: reduce
|
10255 ms |
|
org.apache.spark.JavaAPISuite: sample
|
10176 ms |
|
org.apache.spark.JavaAPISuite: sortBy
|
10264 ms |
|
org.apache.spark.JavaAPISuite: mapsFromPairsToPairs
|
10154 ms |
|
org.apache.spark.JavaAPISuite: flatMap
|
10134 ms |
|
org.apache.spark.JavaAPISuite: cogroup3
|
10793 ms |
|
org.apache.spark.JavaAPISuite: cogroup4
|
10178 ms |
|
org.apache.spark.JavaAPISuite: randomSplit
|
10164 ms |
|
org.apache.spark.JavaAPISuite: persist
|
10267 ms |
|
org.apache.spark.JavaAPISuite: foreach
|
10284 ms |
|
org.apache.spark.JavaAPISuite: hadoopFileCompressed
|
10209 ms |
|
org.apache.spark.JavaAPISuite: accumulators
|
10138 ms |
|
org.apache.spark.JavaAPISuite: textFilesCompressed
|
10196 ms |
|
org.apache.spark.JavaAPISuite: testAsyncActionCancellation
|
10134 ms |
|
org.apache.spark.JavaAPISuite: checkpointAndRestore
|
10158 ms |
|
org.apache.spark.JavaAPISuite: sortByKey
|
10161 ms |
|
org.apache.spark.JavaAPISuite: aggregateByKey
|
10161 ms |
|
org.apache.spark.JavaAPISuite: map
|
10284 ms |
|
org.apache.spark.JavaAPISuite: max
|
10144 ms |
|
org.apache.spark.JavaAPISuite: min
|
10263 ms |
|
org.apache.spark.JavaAPISuite: top
|
10176 ms |
|
org.apache.spark.JavaAPISuite: zip
|
10215 ms |
|
org.apache.spark.JavaAPISuite: fold
|
10272 ms |
|
org.apache.spark.JavaAPISuite: glom
|
10157 ms |
|
org.apache.spark.JavaAPISuite: take
|
10226 ms |
|
org.apache.spark.JavaAPISuite: javaDoubleRDDHistoGram
|
10238 ms |
|
org.apache.spark.JavaAPISuite: collectUnderlyingScalaRDD
|
10164 ms |
|
org.apache.spark.JavaAPISuite: keyBy
|
10146 ms |
|
org.apache.spark.JavaAPISuite: mapPartitionsWithIndex
|
10160 ms |
|
org.apache.spark.JavaAPISuite: sampleByKey
|
10128 ms |
|
org.apache.spark.JavaAPISuite: intersection
|
10257 ms |
|
org.apache.spark.JavaAPISuite: aggregate
|
10254 ms |
|
org.apache.spark.JavaAPISuite: cartesian
|
10165 ms |
|
org.apache.spark.JavaAPISuite: countApproxDistinctByKey
|
10176 ms |
|
org.apache.spark.JavaAPISuite: readWithNewAPIHadoopFile
|
10165 ms |
|
org.apache.spark.JavaAPISuite: testRegisterKryoClasses
|
10190 ms |
|
org.apache.spark.JavaAPISuite: groupBy
|
10159 ms |
|
org.apache.spark.JavaAPISuite: sampleByKeyExact
|
10137 ms |
|
org.apache.spark.JavaAPISuite: mapPartitions
|
10147 ms |
|
org.apache.spark.JavaAPISuite: takeOrdered
|
10209 ms |
|
org.apache.spark.JavaAPISuite: foldByKey
|
10279 ms |
|
org.apache.spark.JavaAPISuite: objectFilesOfInts
|
10147 ms |
|
org.apache.spark.JavaAPISuite: treeAggregate
|
10139 ms |
|
org.apache.spark.JavaAPISuite: approximateResults
|
10248 ms |
|
org.apache.spark.JavaAPISuite: treeReduce
|
10232 ms |
|
org.apache.spark.JavaAPISuite: collectAsMapAndSerialize
|
10291 ms |
|
org.apache.spark.JavaAPISuite: countApproxDistinct
|
10173 ms |
|
org.apache.spark.JavaAPISuite: javaDoubleRDD
|
10146 ms |
|
org.apache.spark.JavaAPISuite: mapOnPairRDD
|
10161 ms |
|
org.apache.spark.JavaAPISuite: testAsyncActionErrorWrapping
|
10177 ms |
|
org.apache.spark.JavaAPISuite: naturalMax
|
10241 ms |
|
org.apache.spark.JavaAPISuite: naturalMin
|
10165 ms |
|
org.apache.spark.JavaAPISuite: sequenceFile
|
10168 ms |
|
org.apache.spark.JavaAPISuite: collectPartitions
|
10256 ms |
|
org.apache.spark.JavaAPISuite: cogroup
|
10136 ms |
|
org.apache.spark.JavaAPISuite: reduceByKey
|
10163 ms |
|
org.apache.spark.JavaAPISuite: repartition
|
10153 ms |
|
org.apache.spark.JavaAPISuite: iterator
|
10248 ms |
|
org.apache.spark.JavaAPISuite: emptyRDD
|
10152 ms |
|
org.apache.spark.JavaAPISuite: zipWithIndex
|
10149 ms |
|
org.apache.spark.JavaAPISuite: foreachPartition
|
10284 ms |
|
org.apache.spark.JavaAPISuite: combineByKey
|
10139 ms |
|
org.apache.spark.JavaAPISuite: takeAsync
|
10540 ms |
|
org.apache.spark.JavaAPISuite: collectAsMapWithIntArrayValues
|
10155 ms |
|
org.apache.spark.JavaAPISuite: objectFilesOfComplexTypes
|
10157 ms |
|
org.apache.spark.JavaAPISuite: zipWithUniqueId
|
10155 ms |
|
org.apache.spark.JavaAPISuite: collectAsync
|
10285 ms |
|
org.apache.spark.JavaAPISuite: foreachAsync
|
10173 ms |
|
org.apache.spark.JavaAPISuite: zipPartitions
|
10151 ms |
|
org.apache.spark.JavaAPISuite: reduceOnJavaDoubleRDD
|
10168 ms |
|
org.apache.spark.JavaAPISuite: isEmpty
|
10236 ms |
|
org.apache.spark.JavaJdbcRDDSuite: testJavaJdbcRDD
|
10154 ms |
|
org.apache.spark.JobCancellationSuite: local mode, FIFO scheduler
|
10108 ms |
|
org.apache.spark.JobCancellationSuite: local mode, fair scheduler
|
10117 ms |
|
org.apache.spark.JobCancellationSuite: cluster mode, FIFO scheduler
|
10124 ms |
|
org.apache.spark.JobCancellationSuite: cluster mode, fair scheduler
|
10187 ms |
|
org.apache.spark.JobCancellationSuite: do not put partially executed partitions into cache
|
10126 ms |
|
org.apache.spark.JobCancellationSuite: job group
|
10202 ms |
|
org.apache.spark.JobCancellationSuite: inherited job group (SPARK-6629)
|
10122 ms |
|
org.apache.spark.JobCancellationSuite: job group with interruption
|
10147 ms |
|
org.apache.spark.JobCancellationSuite: two jobs sharing the same stage
|
10125 ms |
|
org.apache.spark.PartitioningSuite: (It is not a test)
|
10941 ms |
|
org.apache.spark.ShuffleNettySuite: groupByKey without compression
|
10171 ms |
|
org.apache.spark.ShuffleNettySuite: shuffle non-zero block size
|
10130 ms |
|
org.apache.spark.ShuffleNettySuite: shuffle serializer
|
10121 ms |
|
org.apache.spark.ShuffleNettySuite: zero sized blocks
|
10124 ms |
|
org.apache.spark.ShuffleNettySuite: zero sized blocks without kryo
|
11666 ms |
|
org.apache.spark.ShuffleNettySuite: shuffle on mutable pairs
|
10663 ms |
|
org.apache.spark.ShuffleNettySuite: sorting on mutable pairs
|
10598 ms |
|
org.apache.spark.ShuffleNettySuite: cogroup using mutable pairs
|
10108 ms |
|
org.apache.spark.ShuffleNettySuite: subtract mutable pairs
|
10134 ms |
|
org.apache.spark.ShuffleNettySuite: sort with Java non serializable class - Kryo
|
10248 ms |
|
org.apache.spark.ShuffleNettySuite: sort with Java non serializable class - Java
|
10140 ms |
|
org.apache.spark.ShuffleNettySuite: shuffle with different compression settings (SPARK-3426)
|
10118 ms |
|
org.apache.spark.ShuffleNettySuite: [SPARK-4085] rerun map stage if reduce stage cannot find its local shuffle file
|
10151 ms |
|
org.apache.spark.ShuffleNettySuite: metrics for shuffle without aggregation
|
10118 ms |
|
org.apache.spark.ShuffleNettySuite: metrics for shuffle with aggregation
|
10754 ms |
|
org.apache.spark.ShuffleNettySuite: multiple simultaneous attempts for one task (SPARK-8029)
|
10128 ms |
|
org.apache.spark.SortShuffleContextCleanerSuite: (It is not a test)
|
10143 ms |
|
org.apache.spark.SortShuffleSuite: groupByKey without compression
|
10119 ms |
|
org.apache.spark.SortShuffleSuite: shuffle non-zero block size
|
10085 ms |
|
org.apache.spark.SortShuffleSuite: shuffle serializer
|
10074 ms |
|
org.apache.spark.SortShuffleSuite: zero sized blocks
|
10082 ms |
|
org.apache.spark.SortShuffleSuite: zero sized blocks without kryo
|
10135 ms |
|
org.apache.spark.SortShuffleSuite: shuffle on mutable pairs
|
10085 ms |
|
org.apache.spark.SortShuffleSuite: sorting on mutable pairs
|
10100 ms |
|
org.apache.spark.SortShuffleSuite: cogroup using mutable pairs
|
10745 ms |
|
org.apache.spark.SortShuffleSuite: subtract mutable pairs
|
10073 ms |
|
org.apache.spark.SortShuffleSuite: sort with Java non serializable class - Kryo
|
10086 ms |
|
org.apache.spark.SortShuffleSuite: sort with Java non serializable class - Java
|
10111 ms |
|
org.apache.spark.SortShuffleSuite: shuffle with different compression settings (SPARK-3426)
|
10078 ms |
|
org.apache.spark.SortShuffleSuite: [SPARK-4085] rerun map stage if reduce stage cannot find its local shuffle file
|
10078 ms |
|
org.apache.spark.SortShuffleSuite: metrics for shuffle without aggregation
|
10080 ms |
|
org.apache.spark.SortShuffleSuite: metrics for shuffle with aggregation
|
10104 ms |
|
org.apache.spark.SortShuffleSuite: multiple simultaneous attempts for one task (SPARK-8029)
|
10118 ms |
|
org.apache.spark.SortShuffleSuite: SortShuffleManager properly cleans up files for shuffles that use the serialized path
|
10086 ms |
|
org.apache.spark.SortShuffleSuite: SortShuffleManager properly cleans up files for shuffles that use the deserialized path
|
10081 ms |
|
org.apache.spark.SparkConfSuite: creating SparkContext with both master and app name
|
10656 ms |
|
org.apache.spark.SparkConfSuite: SparkContext property overriding
|
10075 ms |
|
org.apache.spark.SparkContextInfoSuite: getPersistentRDDs only returns RDDs that are marked as cached
|
10737 ms |
|
org.apache.spark.SparkContextInfoSuite: getPersistentRDDs returns an immutable map
|
10073 ms |
|
org.apache.spark.SparkContextInfoSuite: getRDDStorageInfo only reports on RDDs that actually persist data
|
10056 ms |
|
org.apache.spark.SparkContextInfoSuite: call sites report correct locations
|
10089 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: bad-master
|
10060 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: local
|
10750 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: local-*
|
12520 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: local-n
|
10053 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: local-*-n-failures
|
10053 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: local-n-failures
|
10060 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: bad-local-n
|
10049 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: bad-local-n-failures
|
10049 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: local-default-parallelism
|
10051 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: simr
|
10052 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: local-cluster
|
10086 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: yarn-cluster
|
10050 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: yarn-standalone
|
10082 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: yarn-client
|
10048 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: mesos fine-grained
|
10051 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: mesos coarse-grained
|
10051 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: mesos with zookeeper
|
10054 ms |
|
org.apache.spark.SparkContextSchedulerCreationSuite: mesos with zookeeper and Master URL starting with zk://
|
10054 ms |
|
org.apache.spark.SparkContextSuite: Only one SparkContext may be active at a time
|
10052 ms |
|
org.apache.spark.SparkContextSuite: Can still construct a new SparkContext after failing to construct a previous one
|
10033 ms |
|
org.apache.spark.SparkContextSuite: Check for multiple SparkContexts can be disabled via undocumented debug option
|
10042 ms |
|
org.apache.spark.SparkContextSuite: Test getOrCreate
|
10033 ms |
|
org.apache.spark.SparkContextSuite: addFile works
|
10041 ms |
|
org.apache.spark.SparkContextSuite: addFile recursive works
|
10033 ms |
|
org.apache.spark.SparkContextSuite: addFile recursive can't add directories by default
|
10032 ms |
|
org.apache.spark.SparkContextSuite: Cancelling job group should not cause SparkContext to shutdown (SPARK-6414)
|
10040 ms |
|
org.apache.spark.SparkContextSuite: Comma separated paths for newAPIHadoopFile/wholeTextFiles/binaryFiles (SPARK-7155)
|
10036 ms |
|
org.apache.spark.SparkContextSuite: Default path for file based RDDs is properly set (SPARK-12517)
|
10040 ms |
|
org.apache.spark.SparkContextSuite: stop() must not throw any exception
|
10054 ms |
|
org.apache.spark.SparkContextSuite: No exception when both num-executors and dynamic allocation set
|
10053 ms |
|
org.apache.spark.StatusTrackerSuite: basic status API usage
|
10046 ms |
|
org.apache.spark.StatusTrackerSuite: getJobIdsForGroup()
|
10057 ms |
|
org.apache.spark.StatusTrackerSuite: getJobIdsForGroup() with takeAsync()
|
10037 ms |
|
org.apache.spark.StatusTrackerSuite: getJobIdsForGroup() with takeAsync() across multiple partitions
|
10070 ms |
|
org.apache.spark.ThreadingSuite: accessing SparkContext form a different thread
|
10115 ms |
|
org.apache.spark.ThreadingSuite: accessing SparkContext form multiple threads
|
10132 ms |
|
org.apache.spark.ThreadingSuite: accessing multi-threaded SparkContext form multiple threads
|
10215 ms |
|
org.apache.spark.ThreadingSuite: parallel job execution
|
10118 ms |
|
org.apache.spark.ThreadingSuite: set local properties in different thread
|
10106 ms |
|
org.apache.spark.ThreadingSuite: set and get local properties in parent-children thread
|
10178 ms |
|
org.apache.spark.ThreadingSuite: mutation in parent local property does not affect child (SPARK-10563)
|
10131 ms |
|
org.apache.spark.UnpersistSuite: unpersist RDD
|
10031 ms |
|
org.apache.spark.api.python.PythonBroadcastSuite: (It is not a test)
|
10118 ms |
|
org.apache.spark.api.python.SerDeUtilSuite: (It is not a test)
|
10068 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Using TorrentBroadcast locally
|
10090 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Accessing TorrentBroadcast variables from multiple threads
|
10106 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Accessing TorrentBroadcast variables in a local cluster
|
10109 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Test Lazy Broadcast variables with TorrentBroadcast
|
10164 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Unpersisting TorrentBroadcast on executors only in local mode
|
10099 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Unpersisting TorrentBroadcast on executors and driver in local mode
|
10096 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Unpersisting TorrentBroadcast on executors only in distributed mode
|
10104 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Unpersisting TorrentBroadcast on executors and driver in distributed mode
|
10103 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Using broadcast after destroy prints callsite
|
10104 ms |
|
org.apache.spark.broadcast.BroadcastSuite: Broadcast variables cannot be created after SparkContext is stopped (SPARK-5065)
|
10107 ms |
|
org.apache.spark.deploy.LogUrlsStandaloneSuite: verify that correct log urls get propagated from workers
|
10046 ms |
|
org.apache.spark.deploy.LogUrlsStandaloneSuite: verify that log urls reflect SPARK_PUBLIC_DNS (SPARK-6175)
|
10056 ms |
|
org.apache.spark.deploy.SparkSubmitSuite: launch simple application with spark-submit
|
15702 ms |
|
org.apache.spark.deploy.SparkSubmitSuite: includes jars passed in through --jars
|
14850 ms |
|
org.apache.spark.deploy.SparkSubmitSuite: includes jars passed in through --packages
|
17447 ms |
|
org.apache.spark.deploy.StandaloneDynamicAllocationSuite: dynamic allocation default behavior
|
10195 ms |
|
org.apache.spark.deploy.StandaloneDynamicAllocationSuite: dynamic allocation with max cores <= cores per worker
|
10118 ms |
|
org.apache.spark.deploy.StandaloneDynamicAllocationSuite: dynamic allocation with max cores > cores per worker
|
10186 ms |
|
org.apache.spark.deploy.StandaloneDynamicAllocationSuite: dynamic allocation with cores per executor
|
10152 ms |
|
org.apache.spark.deploy.StandaloneDynamicAllocationSuite: dynamic allocation with cores per executor AND max cores
|
10185 ms |
|
org.apache.spark.deploy.StandaloneDynamicAllocationSuite: kill the same executor twice (SPARK-9795)
|
10105 ms |
|
org.apache.spark.deploy.StandaloneDynamicAllocationSuite: the pending replacement executors should not be lost (SPARK-10515)
|
10122 ms |
|
org.apache.spark.deploy.StandaloneDynamicAllocationSuite: disable force kill for busy executors (SPARK-9552)
|
10118 ms |
|
org.apache.spark.input.WholeTextFileRecordReaderSuite: (It is not a test)
|
10125 ms |
|
org.apache.spark.launcher.LauncherBackendSuite: local: launcher handle
|
30070 ms |
|
org.apache.spark.launcher.LauncherBackendSuite: standalone/client: launcher handle
|
30089 ms |
|
org.apache.spark.metrics.InputOutputMetricsSuite: (It is not a test)
|
10029 ms |
|
org.apache.spark.rdd.AsyncRDDActionsSuite: (It is not a test)
|
10040 ms |
|
org.apache.spark.rdd.DoubleRDDSuite: (It is not a test)
|
10526 ms |
|
org.apache.spark.rdd.JdbcRDDSuite: basic functionality
|
10103 ms |
|
org.apache.spark.rdd.JdbcRDDSuite: large id overflow
|
10088 ms |
|
org.apache.spark.rdd.LocalCheckpointSuite: (It is not a test)
|
10099 ms |
|
org.apache.spark.rdd.PairRDDFunctionsSuite: (It is not a test)
|
10035 ms |
|
org.apache.spark.rdd.PartitionPruningRDDSuite: (It is not a test)
|
10032 ms |
|
org.apache.spark.rdd.PartitionwiseSampledRDDSuite: (It is not a test)
|
10111 ms |
|
org.apache.spark.rdd.PipedRDDSuite: (It is not a test)
|
10073 ms |
|
org.apache.spark.rdd.RDDOperationScopeSuite: (It is not a test)
|
10230 ms |
|
org.apache.spark.rdd.RDDSuite: (It is not a test)
|
10049 ms |
|
org.apache.spark.rdd.SortingSuite: (It is not a test)
|
10152 ms |
|
org.apache.spark.rdd.ZippedPartitionsSuite: (It is not a test)
|
10117 ms |
|
org.apache.spark.scheduler.AdaptiveSchedulingSuite: simple use of submitMapStage
|
10086 ms |
|
org.apache.spark.scheduler.AdaptiveSchedulingSuite: fetching multiple map output partitions per reduce
|
11711 ms |
|
org.apache.spark.scheduler.AdaptiveSchedulingSuite: fetching all map output partitions in one reduce
|
10086 ms |
|
org.apache.spark.scheduler.AdaptiveSchedulingSuite: more reduce tasks than map output partitions
|
10085 ms |
|
org.apache.spark.scheduler.CoarseGrainedSchedulerBackendSuite: serialized task larger than akka frame size
|
10666 ms |
|
org.apache.spark.scheduler.DAGSchedulerSuite: (It is not a test)
|
10082 ms |
|
org.apache.spark.scheduler.EventLoggingListenerSuite: End-to-end event logging
|
10103 ms |
|
org.apache.spark.scheduler.EventLoggingListenerSuite: End-to-end event logging with compression
|
10151 ms |
|
org.apache.spark.scheduler.OutputCommitCoordinatorIntegrationSuite: (It is not a test)
|
10032 ms |
|
org.apache.spark.scheduler.OutputCommitCoordinatorSuite: (It is not a test)
|
11169 ms |
|
org.apache.spark.scheduler.PoolSuite: FIFO Scheduler Test
|
10092 ms |
|
org.apache.spark.scheduler.PoolSuite: Fair Scheduler Test
|
10104 ms |
|
org.apache.spark.scheduler.PoolSuite: Nested Pool Test
|
10113 ms |
|
org.apache.spark.scheduler.ReplayListenerSuite: End-to-end replay
|
10036 ms |
|
org.apache.spark.scheduler.ReplayListenerSuite: End-to-end replay with compression
|
10037 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: stop in listener
|
10060 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: basic creation of StageInfo
|
10066 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: basic creation of StageInfo with shuffle
|
10042 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: StageInfo with fewer tasks than partitions
|
10043 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: local metrics
|
10039 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: onTaskGettingResult() called when result fetched remotely
|
10396 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: onTaskGettingResult() not called when result sent directly
|
10444 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: onTaskEnd() should be called for all started tasks, even after job has been killed
|
10040 ms |
|
org.apache.spark.scheduler.SparkListenerSuite: extraListeners
|
10053 ms |
|
org.apache.spark.scheduler.SparkListenerWithClusterSuite: (It is not a test)
|
10066 ms |
|
org.apache.spark.scheduler.TaskContextSuite: provide metrics sources
|
10059 ms |
|
org.apache.spark.scheduler.TaskContextSuite: calls TaskCompletionListener after failure
|
10081 ms |
|
org.apache.spark.scheduler.TaskContextSuite: attemptNumber should return attempt number, not task id (SPARK-4014)
|
10085 ms |
|
org.apache.spark.scheduler.TaskResultGetterSuite: handling results smaller than Akka frame size
|
10037 ms |
|
org.apache.spark.scheduler.TaskResultGetterSuite: handling results larger than Akka frame size
|
10034 ms |
|
org.apache.spark.scheduler.TaskResultGetterSuite: task retried if result missing from block manager
|
10040 ms |
|
org.apache.spark.scheduler.TaskResultGetterSuite: failed task deserialized with the correct classloader (SPARK-11195)
|
11437 ms |
|
org.apache.spark.scheduler.TaskSchedulerImplSuite: Scheduler does not always schedule tasks on the same workers
|
10161 ms |
|
org.apache.spark.scheduler.TaskSchedulerImplSuite: Scheduler correctly accounts for multiple CPUs per task
|
10745 ms |
|
org.apache.spark.scheduler.TaskSchedulerImplSuite: Scheduler does not crash when tasks are not serializable
|
10118 ms |
|
org.apache.spark.scheduler.TaskSchedulerImplSuite: refuse to schedule concurrent attempts for the same stage (SPARK-8103)
|
10139 ms |
|
org.apache.spark.scheduler.TaskSchedulerImplSuite: don't schedule more tasks after a taskset is zombie
|
10130 ms |
|
org.apache.spark.scheduler.TaskSchedulerImplSuite: if a zombie attempt finishes, continue scheduling tasks for non-zombie attempts
|
10115 ms |
|
org.apache.spark.scheduler.TaskSchedulerImplSuite: tasks are not re-scheduled while executor loss reason is pending
|
10121 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: TaskSet with no preferences
|
10027 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: multiple offers with no preferences
|
10027 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: skip unsatisfiable locality levels
|
10033 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: basic delay scheduling
|
10034 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: we do not need to delay scheduling when we only have noPref tasks in the queue
|
10035 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: delay scheduling with fallback
|
10037 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: delay scheduling with failed hosts
|
10067 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: task result lost
|
10035 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: repeated failures lead to task set abortion
|
10057 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: executors should be blacklisted after task failure, in spite of locality preferences
|
10041 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: new executors get added and lost
|
10053 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: Executors exit for reason unrelated to currently running tasks
|
10037 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: test RACK_LOCAL tasks
|
10033 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: do not emit warning when serialized task is small
|
10055 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: emit warning when serialized task is large
|
10058 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: Not serializable exception thrown if the task cannot be serialized
|
10037 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: abort the job if total size of results is too large
|
10038 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: speculative and noPref task should be scheduled after node-local
|
10036 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: node-local tasks should be scheduled right away when there are only node-local and no-preference tasks
|
10043 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: SPARK-4939: node-local tasks should be scheduled right after process-local tasks finished
|
10036 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: SPARK-4939: no-pref tasks should be scheduled after process-local tasks finished
|
10063 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: Ensure TaskSetManager is usable after addition of levels
|
10539 ms |
|
org.apache.spark.scheduler.TaskSetManagerSuite: Test that locations with HDFSCacheTaskLocation are treated as PROCESS_LOCAL
|
10508 ms |
|
org.apache.spark.scheduler.cluster.mesos.CoarseMesosSchedulerBackendSuite: (It is not a test)
|
10093 ms |
|
org.apache.spark.serializer.GenericAvroSerializerSuite: (It is not a test)
|
10809 ms |
|
org.apache.spark.serializer.KryoSerializerAutoResetDisabledSuite: (It is not a test)
|
10116 ms |
|
org.apache.spark.serializer.KryoSerializerDistributedSuite: kryo objects are serialised consistently in different processes
|
10306 ms |
|
org.apache.spark.serializer.KryoSerializerResizableOutputSuite: kryo without resizable output buffer should fail on large array
|
10129 ms |
|
org.apache.spark.serializer.KryoSerializerResizableOutputSuite: kryo with resizable output buffer should succeed on large array
|
10094 ms |
|
org.apache.spark.serializer.KryoSerializerSuite: (It is not a test)
|
10108 ms |
|
org.apache.spark.serializer.ProactiveClosureSerializationSuite: (It is not a test)
|
10088 ms |
|
org.apache.spark.shuffle.BlockStoreShuffleReaderSuite: read() releases resources on completion
|
10223 ms |
|
org.apache.spark.shuffle.ShuffleDependencySuite: key, value, and combiner classes correct in shuffle dependency without aggregation
|
10063 ms |
|
org.apache.spark.shuffle.ShuffleDependencySuite: key, value, and combiner classes available in shuffle dependency with aggregation
|
10085 ms |
|
org.apache.spark.shuffle.ShuffleDependencySuite: combineByKey null combiner class tag handled correctly
|
10078 ms |
|
org.apache.spark.storage.FlatmapIteratorSuite: Flatmap Iterator to Disk
|
10044 ms |
|
org.apache.spark.storage.FlatmapIteratorSuite: Flatmap Iterator to Memory
|
10029 ms |
|
org.apache.spark.storage.FlatmapIteratorSuite: Serializer Reset
|
10035 ms |
|
org.apache.spark.ui.UISeleniumSuite: effects of unpersist() / persist() should be reflected
|
10133 ms |
|
org.apache.spark.ui.UISeleniumSuite: failed stages should not appear to be active
|
10679 ms |
|
org.apache.spark.ui.UISeleniumSuite: killEnabled should properly control kill button display
|
10126 ms |
|
org.apache.spark.ui.UISeleniumSuite: jobs page should not display job group name unless some job was submitted in a job group
|
10221 ms |
|
org.apache.spark.ui.UISeleniumSuite: job progress bars should handle stage / task failures
|
10124 ms |
|
org.apache.spark.ui.UISeleniumSuite: job details page should display useful information for stages that haven't started
|
10140 ms |
|
org.apache.spark.ui.UISeleniumSuite: job progress bars / cells reflect skipped stages / tasks
|
10135 ms |
|
org.apache.spark.ui.UISeleniumSuite: stages that aren't run appear as 'skipped stages' after a job finishes
|
10136 ms |
|
org.apache.spark.ui.UISeleniumSuite: jobs with stages that are skipped should show correct link descriptions on all jobs page
|
10871 ms |
|
org.apache.spark.ui.UISeleniumSuite: attaching and detaching a new tab
|
10149 ms |
|
org.apache.spark.ui.UISeleniumSuite: kill stage POST/GET response is correct
|
10146 ms |
|
org.apache.spark.ui.UISeleniumSuite: stage & job retention
|
10138 ms |
|
org.apache.spark.ui.UISeleniumSuite: live UI json application list
|
10231 ms |
|
org.apache.spark.ui.UISeleniumSuite: job stages should have expected dotfile under DAG visualization
|
10157 ms |
|
org.apache.spark.ui.UISuite: verify appUIAddress contains the scheme
|
11183 ms |
|
org.apache.spark.ui.UISuite: verify appUIAddress contains the port
|
10183 ms |
|
org.apache.spark.util.ClosureCleanerSuite: closures inside an object
|
10039 ms |
|
org.apache.spark.util.ClosureCleanerSuite: closures inside a class
|
10114 ms |
|
org.apache.spark.util.ClosureCleanerSuite: closures inside a class with no default constructor
|
10058 ms |
|
org.apache.spark.util.ClosureCleanerSuite: closures that don't use fields of the outer class
|
10077 ms |
|
org.apache.spark.util.ClosureCleanerSuite: nested closures inside an object
|
10074 ms |
|
org.apache.spark.util.ClosureCleanerSuite: nested closures inside a class
|
10095 ms |
|
org.apache.spark.util.ClosureCleanerSuite: toplevel return statements in closures are identified at cleaning time
|
10083 ms |
|
org.apache.spark.util.ClosureCleanerSuite: return statements from named functions nested in closures don't raise exceptions
|
10074 ms |
|
org.apache.spark.util.ClosureCleanerSuite: user provided closures are actually cleaned
|
10083 ms |
|
org.apache.spark.util.ClosureCleanerSuite2: (It is not a test)
|
10059 ms |
|
org.apache.spark.util.MutableURLClassLoaderSuite: driver sets context class loader in local mode
|
10212 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: single insert insert
|
14278 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: multiple insert
|
12115 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: insert with collision
|
10119 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: ordering
|
11150 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: null keys and values
|
10130 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: simple aggregator
|
10113 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: simple cogroup
|
10118 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: spilling
|
10125 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: spilling with compression
|
10141 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: spilling with hash collisions
|
11066 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: spilling with many hash collisions
|
10803 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: MaxValue key
|
10130 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: spilling with null keys and values
|
10129 ms |
|
org.apache.spark.util.collection.ExternalAppendOnlyMapSuite: external aggregation updates peak execution memory
|
10126 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: empty data stream with kryo ser
|
10056 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: empty data stream with java ser
|
10068 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: few elements per partition with kryo ser
|
10064 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: few elements per partition with java ser
|
10062 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: empty partitions with spilling with kryo ser
|
10104 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: empty partitions with spilling with java ser
|
10080 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: spilling in local cluster with kryo ser
|
10070 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: spilling in local cluster with java ser
|
10063 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: spilling in local cluster with many reduce tasks with kryo ser
|
10066 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: spilling in local cluster with many reduce tasks with java ser
|
10060 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: cleanup of intermediate files in sorter
|
10096 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: cleanup of intermediate files in sorter with failures
|
10065 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: cleanup of intermediate files in shuffle
|
10071 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: cleanup of intermediate files in shuffle with failures
|
10065 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: no sorting or partial aggregation with kryo ser
|
10070 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: no sorting or partial aggregation with java ser
|
10061 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: no sorting or partial aggregation with spilling with kryo ser
|
10110 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: no sorting or partial aggregation with spilling with java ser
|
10066 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: sorting, no partial aggregation with kryo ser
|
10069 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: sorting, no partial aggregation with java ser
|
10111 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: sorting, no partial aggregation with spilling with kryo ser
|
10078 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: sorting, no partial aggregation with spilling with java ser
|
10125 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: partial aggregation, no sorting with kryo ser
|
10068 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: partial aggregation, no sorting with java ser
|
10065 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: partial aggregation, no sorting with spilling with kryo ser
|
11333 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: partial aggregation, no sorting with spilling with java ser
|
10123 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: partial aggregation and sorting with kryo ser
|
10069 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: partial aggregation and sorting with java ser
|
10081 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: partial aggregation and sorting with spilling with kryo ser
|
10093 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: partial aggregation and sorting with spilling with java ser
|
10066 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: sort without breaking sorting contracts with kryo ser
|
10073 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: sort without breaking sorting contracts with java ser
|
11028 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: spilling with hash collisions
|
10067 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: spilling with many hash collisions
|
10066 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: MaxValue key
|
10077 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: spilling with null keys and values
|
10075 ms |
|
org.apache.spark.util.collection.ExternalSorterSuite: sorting updates peak execution memory
|
10115 ms |
|
org.apache.spark.streaming.kafka.DirectKafkaStreamSuite: basic stream receiving with multiple topics and smallest starting offset
|
10831 ms |
|
org.apache.spark.streaming.kafka.DirectKafkaStreamSuite: receiving from largest starting offset
|
10297 ms |
|
org.apache.spark.streaming.kafka.DirectKafkaStreamSuite: creating stream by offset
|
10333 ms |
|
org.apache.spark.streaming.kafka.DirectKafkaStreamSuite: offset recovery
|
10200 ms |
|
org.apache.spark.streaming.kafka.DirectKafkaStreamSuite: Direct Kafka stream report input information
|
10395 ms |
|
org.apache.spark.streaming.kafka.DirectKafkaStreamSuite: using rate controller
|
10244 ms |
|
org.apache.spark.streaming.kafka.JavaDirectKafkaStreamSuite: testKafkaStream
|
10842 ms |
|
org.apache.spark.streaming.kafka.JavaKafkaRDDSuite: testKafkaRDD
|
10752 ms |
|
org.apache.spark.streaming.kafka.JavaKafkaStreamSuite: testKafkaStream
|
10670 ms |
|
org.apache.spark.streaming.kafka.KafkaRDDSuite: (It is not a test)
|
10059 ms |
|
org.apache.spark.streaming.kafka.KafkaStreamSuite: Kafka input stream
|
10073 ms |
|
org.apache.spark.streaming.kafka.ReliableKafkaStreamSuite: (It is not a test)
|
12700 ms |
|
org.apache.spark.streaming.mqtt.JavaMQTTStreamSuite: testMQTTStream
|
10086 ms |
|
org.apache.spark.streaming.mqtt.MQTTStreamSuite: (It is not a test)
|
11616 ms |
|
org.apache.spark.streaming.twitter.JavaTwitterStreamSuite: testTwitterStream
|
10083 ms |
|
org.apache.spark.streaming.twitter.TwitterStreamSuite: twitter input stream
|
11553 ms |
|
org.apache.spark.streaming.zeromq.JavaZeroMQStreamSuite: testZeroMQStream
|
10087 ms |
|
org.apache.spark.streaming.zeromq.ZeroMQStreamSuite: zeromq input stream
|
11579 ms |
|
org.apache.spark.streaming.kinesis.JavaKinesisStreamSuite: testCustomHandler
|
10086 ms |
|
org.apache.spark.streaming.kinesis.JavaKinesisStreamSuite: testKinesisStream
|
10064 ms |
|
org.apache.spark.streaming.kinesis.WithAggregationKinesisStreamSuite: (It is not a test)
|
11173 ms |
|
org.apache.spark.streaming.kinesis.WithoutAggregationKinesisStreamSuite: (It is not a test)
|
10078 ms |
|
org.apache.spark.graphx.EdgeRDDSuite: cache, getStorageLevel
|
10057 ms |
|
org.apache.spark.graphx.GraphLoaderSuite: edgeListFile
|
10043 ms |
|
org.apache.spark.graphx.GraphOpsSuite: joinVertices
|
10060 ms |
|
org.apache.spark.graphx.GraphOpsSuite: collectNeighborIds
|
10076 ms |
|
org.apache.spark.graphx.GraphOpsSuite: filter
|
10105 ms |
|
org.apache.spark.graphx.GraphOpsSuite: convertToCanonicalEdges
|
11040 ms |
|
org.apache.spark.graphx.GraphOpsSuite: collectEdgesCycleDirectionOut
|
10039 ms |
|
org.apache.spark.graphx.GraphOpsSuite: collectEdgesCycleDirectionIn
|
10043 ms |
|
org.apache.spark.graphx.GraphOpsSuite: collectEdgesCycleDirectionEither
|
10066 ms |
|
org.apache.spark.graphx.GraphOpsSuite: collectEdgesChainDirectionOut
|
10044 ms |
|
org.apache.spark.graphx.GraphOpsSuite: collectEdgesChainDirectionIn
|
10042 ms |
|
org.apache.spark.graphx.GraphOpsSuite: collectEdgesChainDirectionEither
|
10043 ms |
|
org.apache.spark.graphx.GraphSuite: fromEdgeTuples
|
10036 ms |
|
org.apache.spark.graphx.GraphSuite: fromEdges
|
10038 ms |
|
org.apache.spark.graphx.GraphSuite: apply
|
10040 ms |
|
org.apache.spark.graphx.GraphSuite: triplets
|
10036 ms |
|
org.apache.spark.graphx.GraphSuite: partitionBy
|
10038 ms |
|
org.apache.spark.graphx.GraphSuite: mapVertices
|
10041 ms |
|
org.apache.spark.graphx.GraphSuite: mapVertices changing type with same erased type
|
10039 ms |
|
org.apache.spark.graphx.GraphSuite: mapEdges
|
10051 ms |
|
org.apache.spark.graphx.GraphSuite: mapTriplets
|
10053 ms |
|
org.apache.spark.graphx.GraphSuite: reverse
|
10071 ms |
|
org.apache.spark.graphx.GraphSuite: reverse with join elimination
|
10040 ms |
|
org.apache.spark.graphx.GraphSuite: subgraph
|
10045 ms |
|
org.apache.spark.graphx.GraphSuite: mask
|
10041 ms |
|
org.apache.spark.graphx.GraphSuite: groupEdges
|
10049 ms |
|
org.apache.spark.graphx.GraphSuite: mapReduceTriplets
|
10127 ms |
|
org.apache.spark.graphx.GraphSuite: aggregateMessages
|
10059 ms |
|
org.apache.spark.graphx.GraphSuite: outerJoinVertices
|
10047 ms |
|
org.apache.spark.graphx.GraphSuite: more edge partitions than vertex partitions
|
10052 ms |
|
org.apache.spark.graphx.GraphSuite: checkpoint
|
10058 ms |
|
org.apache.spark.graphx.GraphSuite: cache, getStorageLevel
|
10045 ms |
|
org.apache.spark.graphx.GraphSuite: non-default number of edge partitions
|
10056 ms |
|
org.apache.spark.graphx.PregelSuite: 1 iteration
|
10043 ms |
|
org.apache.spark.graphx.PregelSuite: chain propagation
|
10042 ms |
|
org.apache.spark.graphx.VertexRDDSuite: filter
|
10041 ms |
|
org.apache.spark.graphx.VertexRDDSuite: mapValues
|
10044 ms |
|
org.apache.spark.graphx.VertexRDDSuite: minus
|
10044 ms |
|
org.apache.spark.graphx.VertexRDDSuite: minus with RDD[(VertexId, VD)]
|
10038 ms |
|
org.apache.spark.graphx.VertexRDDSuite: minus with non-equal number of partitions
|
10038 ms |
|
org.apache.spark.graphx.VertexRDDSuite: diff
|
10037 ms |
|
org.apache.spark.graphx.VertexRDDSuite: diff with RDD[(VertexId, VD)]
|
10052 ms |
|
org.apache.spark.graphx.VertexRDDSuite: diff vertices with non-equal number of partitions
|
10063 ms |
|
org.apache.spark.graphx.VertexRDDSuite: leftJoin
|
10041 ms |
|
org.apache.spark.graphx.VertexRDDSuite: leftJoin vertices with non-equal number of partitions
|
10069 ms |
|
org.apache.spark.graphx.VertexRDDSuite: innerJoin
|
10046 ms |
|
org.apache.spark.graphx.VertexRDDSuite: innerJoin vertices with the non-equal number of partitions
|
10055 ms |
|
org.apache.spark.graphx.VertexRDDSuite: aggregateUsingIndex
|
10043 ms |
|
org.apache.spark.graphx.VertexRDDSuite: mergeFunc
|
10038 ms |
|
org.apache.spark.graphx.VertexRDDSuite: cache, getStorageLevel
|
10050 ms |
|
org.apache.spark.graphx.lib.ConnectedComponentsSuite: Grid Connected Components
|
10042 ms |
|
org.apache.spark.graphx.lib.ConnectedComponentsSuite: Reverse Grid Connected Components
|
10055 ms |
|
org.apache.spark.graphx.lib.ConnectedComponentsSuite: Chain Connected Components
|
10066 ms |
|
org.apache.spark.graphx.lib.ConnectedComponentsSuite: Reverse Chain Connected Components
|
10047 ms |
|
org.apache.spark.graphx.lib.ConnectedComponentsSuite: Connected Components on a Toy Connected Graph
|
10047 ms |
|
org.apache.spark.graphx.lib.LabelPropagationSuite: Label Propagation
|
10070 ms |
|
org.apache.spark.graphx.lib.PageRankSuite: Star PageRank
|
10108 ms |
|
org.apache.spark.graphx.lib.PageRankSuite: Star PersonalPageRank
|
10058 ms |
|
org.apache.spark.graphx.lib.PageRankSuite: Grid PageRank
|
10051 ms |
|
org.apache.spark.graphx.lib.PageRankSuite: Chain PageRank
|
10051 ms |
|
org.apache.spark.graphx.lib.PageRankSuite: Chain PersonalizedPageRank
|
10047 ms |
|
org.apache.spark.graphx.lib.SVDPlusPlusSuite: Test SVD++ with mean square error on training set
|
10036 ms |
|
org.apache.spark.graphx.lib.ShortestPathsSuite: Shortest Path Computations
|
10048 ms |
|
org.apache.spark.graphx.lib.StronglyConnectedComponentsSuite: Island Strongly Connected Components
|
10037 ms |
|
org.apache.spark.graphx.lib.StronglyConnectedComponentsSuite: Cycle Strongly Connected Components
|
10037 ms |
|
org.apache.spark.graphx.lib.StronglyConnectedComponentsSuite: 2 Cycle Strongly Connected Components
|
10037 ms |
|
org.apache.spark.graphx.lib.TriangleCountSuite: Count a single triangle
|
11551 ms |
|
org.apache.spark.graphx.lib.TriangleCountSuite: Count two triangles
|
10093 ms |
|
org.apache.spark.graphx.lib.TriangleCountSuite: Count two triangles with bi-directed edges
|
10068 ms |
|
org.apache.spark.graphx.lib.TriangleCountSuite: Count a single triangle with duplicate edges
|
10059 ms |
|
org.apache.spark.graphx.util.GraphGeneratorsSuite: logNormalGraph
|
10062 ms |
|
org.apache.spark.graphx.util.GraphGeneratorsSuite: rmatGraph numEdges upper bound
|
10043 ms |
|
org.apache.spark.ml.PipelineSuite: (It is not a test)
|
10146 ms |
|
org.apache.spark.ml.ann.ANNSuite: (It is not a test)
|
11165 ms |
|
org.apache.spark.ml.classification.DecisionTreeClassifierSuite: (It is not a test)
|
10060 ms |
|
org.apache.spark.ml.classification.GBTClassifierSuite: (It is not a test)
|
10093 ms |
|
org.apache.spark.ml.classification.JavaDecisionTreeClassifierSuite: runDT
|
10057 ms |
|
org.apache.spark.ml.classification.LogisticRegressionSuite: (It is not a test)
|
10510 ms |
|
org.apache.spark.ml.classification.MultilayerPerceptronClassifierSuite: (It is not a test)
|
10049 ms |
|
org.apache.spark.ml.classification.NaiveBayesSuite: (It is not a test)
|
10043 ms |
|
org.apache.spark.ml.classification.OneVsRestSuite: (It is not a test)
|
10054 ms |
|
org.apache.spark.ml.classification.RandomForestClassifierSuite: (It is not a test)
|
10064 ms |
|
org.apache.spark.ml.clustering.KMeansSuite: (It is not a test)
|
10062 ms |
|
org.apache.spark.ml.clustering.LDASuite: (It is not a test)
|
10067 ms |
|
org.apache.spark.ml.evaluation.BinaryClassificationEvaluatorSuite: (It is not a test)
|
10047 ms |
|
org.apache.spark.ml.evaluation.MulticlassClassificationEvaluatorSuite: (It is not a test)
|
10055 ms |
|
org.apache.spark.ml.evaluation.RegressionEvaluatorSuite: (It is not a test)
|
10048 ms |
|
org.apache.spark.ml.feature.BinarizerSuite: (It is not a test)
|
10051 ms |
|
org.apache.spark.ml.feature.BucketizerSuite: (It is not a test)
|
10049 ms |
|
org.apache.spark.ml.feature.ChiSqSelectorSuite: (It is not a test)
|
10049 ms |
|
org.apache.spark.ml.feature.CountVectorizerSuite: (It is not a test)
|
10053 ms |
|
org.apache.spark.ml.feature.DCTSuite: (It is not a test)
|
10056 ms |
|
org.apache.spark.ml.feature.HashingTFSuite: (It is not a test)
|
10059 ms |
|
org.apache.spark.ml.feature.IDFSuite: (It is not a test)
|
10054 ms |
|
org.apache.spark.ml.feature.InteractionSuite: (It is not a test)
|
10075 ms |
|
org.apache.spark.ml.feature.JavaVectorSlicerSuite: vectorSlice
|
10072 ms |
|
org.apache.spark.ml.feature.MinMaxScalerSuite: (It is not a test)
|
10045 ms |
|
org.apache.spark.ml.feature.NGramSuite: (It is not a test)
|
10045 ms |
|
org.apache.spark.ml.feature.NormalizerSuite: (It is not a test)
|
10046 ms |
|
org.apache.spark.ml.feature.OneHotEncoderSuite: (It is not a test)
|
10052 ms |
|
org.apache.spark.ml.feature.PCASuite: (It is not a test)
|
10065 ms |
|
org.apache.spark.ml.feature.PolynomialExpansionSuite: (It is not a test)
|
10697 ms |
|
org.apache.spark.ml.feature.QuantileDiscretizerSuite: (It is not a test)
|
10061 ms |
|
org.apache.spark.ml.feature.RFormulaSuite: (It is not a test)
|
10054 ms |
|
org.apache.spark.ml.feature.RegexTokenizerSuite: (It is not a test)
|
10057 ms |
|
org.apache.spark.ml.feature.SQLTransformerSuite: (It is not a test)
|
10063 ms |
|
org.apache.spark.ml.feature.StandardScalerSuite: (It is not a test)
|
10043 ms |
|
org.apache.spark.ml.feature.StopWordsRemoverSuite: (It is not a test)
|
10055 ms |
|
org.apache.spark.ml.feature.StringIndexerSuite: (It is not a test)
|
10059 ms |
|
org.apache.spark.ml.feature.TokenizerSuite: (It is not a test)
|
10043 ms |
|
org.apache.spark.ml.feature.VectorAssemblerSuite: (It is not a test)
|
10079 ms |
|
org.apache.spark.ml.feature.VectorIndexerSuite: (It is not a test)
|
10051 ms |
|
org.apache.spark.ml.feature.VectorSlicerSuite: (It is not a test)
|
11457 ms |
|
org.apache.spark.ml.feature.Word2VecSuite: (It is not a test)
|
10062 ms |
|
org.apache.spark.ml.optim.WeightedLeastSquaresSuite: (It is not a test)
|
10037 ms |
|
org.apache.spark.ml.recommendation.ALSSuite: (It is not a test)
|
10044 ms |
|
org.apache.spark.ml.regression.AFTSurvivalRegressionSuite: (It is not a test)
|
10037 ms |
|
org.apache.spark.ml.regression.DecisionTreeRegressorSuite: (It is not a test)
|
10042 ms |
|
org.apache.spark.ml.regression.GBTRegressorSuite: (It is not a test)
|
10056 ms |
|
org.apache.spark.ml.regression.IsotonicRegressionSuite: (It is not a test)
|
10040 ms |
|
org.apache.spark.ml.regression.LinearRegressionSuite: (It is not a test)
|
10041 ms |
|
org.apache.spark.ml.regression.RandomForestRegressorSuite: (It is not a test)
|
10088 ms |
|
org.apache.spark.ml.source.libsvm.LibSVMRelationSuite: (It is not a test)
|
10054 ms |
|
org.apache.spark.ml.tree.impl.RandomForestSuite: (It is not a test)
|
10129 ms |
|
org.apache.spark.ml.tuning.CrossValidatorSuite: (It is not a test)
|
10871 ms |
|
org.apache.spark.ml.tuning.TrainValidationSplitSuite: (It is not a test)
|
10055 ms |
|
org.apache.spark.ml.util.DefaultReadWriteSuite: (It is not a test)
|
10088 ms |
|
org.apache.spark.ml.util.StopwatchSuite: (It is not a test)
|
10066 ms |
|
org.apache.spark.mllib.classification.LogisticRegressionClusterSuite: (It is not a test)
|
10043 ms |
|
org.apache.spark.mllib.classification.LogisticRegressionSuite: (It is not a test)
|
10041 ms |
|
org.apache.spark.mllib.classification.NaiveBayesClusterSuite: (It is not a test)
|
10044 ms |
|
org.apache.spark.mllib.classification.NaiveBayesSuite: (It is not a test)
|
10044 ms |
|
org.apache.spark.mllib.classification.SVMClusterSuite: (It is not a test)
|
10742 ms |
|
org.apache.spark.mllib.classification.SVMSuite: (It is not a test)
|
10058 ms |
|
org.apache.spark.mllib.classification.StreamingLogisticRegressionSuite: parameter accuracy
|
10184 ms |
|
org.apache.spark.mllib.classification.StreamingLogisticRegressionSuite: parameter convergence
|
10064 ms |
|
org.apache.spark.mllib.classification.StreamingLogisticRegressionSuite: predictions
|
10072 ms |
|
org.apache.spark.mllib.classification.StreamingLogisticRegressionSuite: training and prediction
|
11177 ms |
|
org.apache.spark.mllib.classification.StreamingLogisticRegressionSuite: handling empty RDDs in a stream
|
10070 ms |
|
org.apache.spark.mllib.clustering.BisectingKMeansSuite: (It is not a test)
|
10059 ms |
|
org.apache.spark.mllib.clustering.GaussianMixtureSuite: (It is not a test)
|
10064 ms |
|
org.apache.spark.mllib.clustering.KMeansClusterSuite: (It is not a test)
|
10064 ms |
|
org.apache.spark.mllib.clustering.KMeansSuite: (It is not a test)
|
10060 ms |
|
org.apache.spark.mllib.clustering.LDASuite: (It is not a test)
|
10044 ms |
|
org.apache.spark.mllib.clustering.PowerIterationClusteringSuite: (It is not a test)
|
10063 ms |
|
org.apache.spark.mllib.clustering.StreamingKMeansSuite: accuracy for single center and equivalence to grand average
|
10063 ms |
|
org.apache.spark.mllib.clustering.StreamingKMeansSuite: accuracy for two centers
|
10070 ms |
|
org.apache.spark.mllib.clustering.StreamingKMeansSuite: detecting dying clusters
|
10059 ms |
|
org.apache.spark.mllib.evaluation.AreaUnderCurveSuite: (It is not a test)
|
10057 ms |
|
org.apache.spark.mllib.evaluation.BinaryClassificationMetricsSuite: (It is not a test)
|
10043 ms |
|
org.apache.spark.mllib.evaluation.MulticlassMetricsSuite: (It is not a test)
|
10080 ms |
|
org.apache.spark.mllib.evaluation.MultilabelMetricsSuite: (It is not a test)
|
10042 ms |
|
org.apache.spark.mllib.evaluation.RankingMetricsSuite: (It is not a test)
|
10045 ms |
|
org.apache.spark.mllib.evaluation.RegressionMetricsSuite: (It is not a test)
|
10055 ms |
|
org.apache.spark.mllib.feature.ChiSqSelectorSuite: (It is not a test)
|
10047 ms |
|
org.apache.spark.mllib.feature.ElementwiseProductSuite: (It is not a test)
|
10068 ms |
|
org.apache.spark.mllib.feature.HashingTFSuite: (It is not a test)
|
10048 ms |
|
org.apache.spark.mllib.feature.IDFSuite: (It is not a test)
|
10050 ms |
|
org.apache.spark.mllib.feature.NormalizerSuite: (It is not a test)
|
10066 ms |
|
org.apache.spark.mllib.feature.PCASuite: (It is not a test)
|
10040 ms |
|
org.apache.spark.mllib.feature.StandardScalerSuite: (It is not a test)
|
10045 ms |
|
org.apache.spark.mllib.feature.Word2VecSuite: (It is not a test)
|
10121 ms |
|
org.apache.spark.mllib.fpm.AssociationRulesSuite: (It is not a test)
|
10062 ms |
|
org.apache.spark.mllib.fpm.FPGrowthSuite: (It is not a test)
|
10058 ms |
|
org.apache.spark.mllib.fpm.FPTreeSuite: (It is not a test)
|
10050 ms |
|
org.apache.spark.mllib.fpm.PrefixSpanSuite: (It is not a test)
|
10054 ms |
|
org.apache.spark.mllib.impl.PeriodicGraphCheckpointerSuite: (It is not a test)
|
10056 ms |
|
org.apache.spark.mllib.impl.PeriodicRDDCheckpointerSuite: (It is not a test)
|
11429 ms |
|
org.apache.spark.mllib.linalg.distributed.BlockMatrixSuite: (It is not a test)
|
12823 ms |
|
org.apache.spark.mllib.linalg.distributed.CoordinateMatrixSuite: (It is not a test)
|
10049 ms |
|
org.apache.spark.mllib.linalg.distributed.IndexedRowMatrixSuite: (It is not a test)
|
10046 ms |
|
org.apache.spark.mllib.linalg.distributed.RowMatrixClusterSuite: (It is not a test)
|
10152 ms |
|
org.apache.spark.mllib.linalg.distributed.RowMatrixSuite: (It is not a test)
|
10058 ms |
|
org.apache.spark.mllib.optimization.GradientDescentClusterSuite: (It is not a test)
|
10052 ms |
|
org.apache.spark.mllib.optimization.GradientDescentSuite: (It is not a test)
|
10053 ms |
|
org.apache.spark.mllib.optimization.LBFGSClusterSuite: (It is not a test)
|
10050 ms |
|
org.apache.spark.mllib.optimization.LBFGSSuite: (It is not a test)
|
10050 ms |
|
org.apache.spark.mllib.random.RandomRDDsSuite: (It is not a test)
|
10039 ms |
|
org.apache.spark.mllib.rdd.MLPairRDDFunctionsSuite: (It is not a test)
|
10044 ms |
|
org.apache.spark.mllib.rdd.RDDFunctionsSuite: (It is not a test)
|
10044 ms |
|
org.apache.spark.mllib.recommendation.ALSSuite: (It is not a test)
|
10054 ms |
|
org.apache.spark.mllib.recommendation.MatrixFactorizationModelSuite: (It is not a test)
|
10057 ms |
|
org.apache.spark.mllib.regression.IsotonicRegressionSuite: (It is not a test)
|
10051 ms |
|
org.apache.spark.mllib.regression.JavaIsotonicRegressionSuite: testIsotonicRegressionJavaRDD
|
10080 ms |
|
org.apache.spark.mllib.regression.JavaIsotonicRegressionSuite: testIsotonicRegressionPredictionsJavaRDD
|
10063 ms |
|
org.apache.spark.mllib.regression.LassoClusterSuite: (It is not a test)
|
10066 ms |
|
org.apache.spark.mllib.regression.LassoSuite: (It is not a test)
|
10042 ms |
|
org.apache.spark.mllib.regression.LinearRegressionClusterSuite: (It is not a test)
|
10062 ms |
|
org.apache.spark.mllib.regression.LinearRegressionSuite: (It is not a test)
|
10072 ms |
|
org.apache.spark.mllib.regression.RidgeRegressionClusterSuite: (It is not a test)
|
10055 ms |
|
org.apache.spark.mllib.regression.RidgeRegressionSuite: (It is not a test)
|
10061 ms |
|
org.apache.spark.mllib.regression.StreamingLinearRegressionSuite: parameter accuracy
|
10085 ms |
|
org.apache.spark.mllib.regression.StreamingLinearRegressionSuite: parameter convergence
|
10059 ms |
|
org.apache.spark.mllib.regression.StreamingLinearRegressionSuite: predictions
|
10066 ms |
|
org.apache.spark.mllib.regression.StreamingLinearRegressionSuite: training and prediction
|
10053 ms |
|
org.apache.spark.mllib.regression.StreamingLinearRegressionSuite: handling empty RDDs in a stream
|
10073 ms |
|
org.apache.spark.mllib.stat.CorrelationSuite: (It is not a test)
|
10064 ms |
|
org.apache.spark.mllib.stat.HypothesisTestSuite: (It is not a test)
|
10071 ms |
|
org.apache.spark.mllib.stat.KernelDensitySuite: (It is not a test)
|
10044 ms |
|
org.apache.spark.mllib.stat.StreamingTestSuite: accuracy for null hypothesis using welch t-test
|
10181 ms |
|
org.apache.spark.mllib.stat.StreamingTestSuite: accuracy for alternative hypothesis using welch t-test
|
10050 ms |
|
org.apache.spark.mllib.stat.StreamingTestSuite: accuracy for null hypothesis using student t-test
|
10056 ms |
|
org.apache.spark.mllib.stat.StreamingTestSuite: accuracy for alternative hypothesis using student t-test
|
10045 ms |
|
org.apache.spark.mllib.stat.StreamingTestSuite: batches within same test window are grouped
|
10064 ms |
|
org.apache.spark.mllib.stat.StreamingTestSuite: entries in peace period are dropped
|
10072 ms |
|
org.apache.spark.mllib.stat.StreamingTestSuite: null hypothesis when only data from one group is present
|
13689 ms |
|
org.apache.spark.mllib.stat.distribution.MultivariateGaussianSuite: (It is not a test)
|
10054 ms |
|
org.apache.spark.mllib.tree.DecisionTreeSuite: (It is not a test)
|
10051 ms |
|
org.apache.spark.mllib.tree.GradientBoostedTreesSuite: (It is not a test)
|
10044 ms |
|
org.apache.spark.mllib.tree.RandomForestSuite: (It is not a test)
|
10043 ms |
|
org.apache.spark.mllib.tree.impl.BaggedPointSuite: (It is not a test)
|
10058 ms |
|
org.apache.spark.mllib.util.MLUtilsSuite: (It is not a test)
|
10037 ms |
|
org.apache.spark.repl.ReplSuite: propagation of local properties
|
11164 ms |
|
org.apache.spark.repl.ReplSuite: simple foreach with accumulator
|
15935 ms |
|
org.apache.spark.repl.ReplSuite: external vars
|
13460 ms |
|
org.apache.spark.repl.ReplSuite: external classes
|
12179 ms |
|
org.apache.spark.repl.ReplSuite: external functions
|
12111 ms |
|
org.apache.spark.repl.ReplSuite: external functions that access vars
|
12898 ms |
|
org.apache.spark.repl.ReplSuite: broadcast vars
|
11750 ms |
|
org.apache.spark.repl.ReplSuite: interacting with files
|
11562 ms |
|
org.apache.spark.repl.ReplSuite: local-cluster mode
|
11789 ms |
|
org.apache.spark.repl.ReplSuite: SPARK-1199 two instances of same class don't type check
|
12170 ms |
|
org.apache.spark.repl.ReplSuite: SPARK-2452 compound statements
|
11593 ms |
|
org.apache.spark.repl.ReplSuite: _
|
11559 ms |
|
org.apache.spark.repl.ReplSuite: SPARK-8461 SQL with codegen
|
13535 ms |
|
org.apache.spark.repl.ReplSuite: Datasets and encoders
|
13942 ms |
|
org.apache.spark.repl.ReplSuite: SPARK-2632 importing a method from non serializable class and not using it
|
12264 ms |
|
org.apache.spark.repl.ReplSuite: Datasets agg type-inference
|
12405 ms |
|
org.apache.spark.repl.ReplSuite: collecting objects of class defined in repl
|
13244 ms |
|
org.apache.spark.repl.ReplSuite: collecting objects of class defined in repl - shuffling
|
12229 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: map
|
10047 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: flatMap
|
10038 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: filter
|
10163 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: glom
|
10060 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: mapPartitions
|
10047 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: repartition (more partitions)
|
10047 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: repartition (fewer partitions)
|
11814 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: groupByKey
|
10049 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: reduceByKey
|
10052 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: reduce
|
10055 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: count
|
10044 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: countByValue
|
10051 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: mapValues
|
10051 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: flatMapValues
|
10049 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: union
|
10049 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: union with input stream return None
|
10069 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: transform
|
10070 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: transform with NULL
|
12118 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: transform with input stream return None
|
10087 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: transformWith
|
10057 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: transformWith with input stream return None
|
10056 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: cogroup
|
10104 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: join
|
10052 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: leftOuterJoin
|
10056 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: rightOuterJoin
|
10065 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: fullOuterJoin
|
10052 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: updateStateByKey
|
11043 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: updateStateByKey - simple with initial value RDD
|
10064 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: updateStateByKey - with initial value RDD
|
10056 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: updateStateByKey - object lifecycle
|
10091 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: slice
|
10050 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: slice - has not been initialized
|
10058 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: rdd cleanup - map and window
|
10056 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: rdd cleanup - updateStateByKey
|
10052 ms |
|
org.apache.spark.streaming.BasicOperationsSuite: rdd cleanup - input blocks and persisted RDDs
|
10088 ms |
|
org.apache.spark.streaming.CheckpointSuite: basic rdd checkpoints + dstream graph checkpoint recovery
|
10111 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery of conf through checkpoints
|
10059 ms |
|
org.apache.spark.streaming.CheckpointSuite: [host|port] from checkpoint
|
10070 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery with map and reduceByKey operations
|
10055 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery with invertible reduceByKeyAndWindow operation
|
10125 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery with saveAsHadoopFiles operation
|
10600 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery with saveAsNewAPIHadoopFiles operation
|
10061 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery with saveAsHadoopFile inside transform operation
|
10053 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery with updateStateByKey operation
|
10069 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery maintains rate controller
|
10057 ms |
|
org.apache.spark.streaming.CheckpointSuite: recovery with file input stream
|
10065 ms |
|
org.apache.spark.streaming.CheckpointSuite: restore invoking times
|
10059 ms |
|
org.apache.spark.streaming.DStreamClosureSuite: (It is not a test)
|
10042 ms |
|
org.apache.spark.streaming.DStreamScopeSuite: (It is not a test)
|
10049 ms |
|
org.apache.spark.streaming.FailureSuite: multiple failures with map
|
10106 ms |
|
org.apache.spark.streaming.FailureSuite: multiple failures with updateStateByKey
|
10065 ms |
|
org.apache.spark.streaming.InputStreamsSuite: socket input stream
|
10071 ms |
|
org.apache.spark.streaming.InputStreamsSuite: socket input stream - no block in a batch
|
10054 ms |
|
org.apache.spark.streaming.InputStreamsSuite: binary records stream
|
10058 ms |
|
org.apache.spark.streaming.InputStreamsSuite: file input stream - newFilesOnly = true
|
10068 ms |
|
org.apache.spark.streaming.InputStreamsSuite: file input stream - newFilesOnly = false
|
10052 ms |
|
org.apache.spark.streaming.InputStreamsSuite: multi-thread receiver
|
10053 ms |
|
org.apache.spark.streaming.InputStreamsSuite: queue input stream - oneAtATime = true
|
10043 ms |
|
org.apache.spark.streaming.InputStreamsSuite: queue input stream - oneAtATime = false
|
10042 ms |
|
org.apache.spark.streaming.InputStreamsSuite: test track the number of input stream
|
10040 ms |
|
org.apache.spark.streaming.JavaAPISuite: testStreamingContextTransform
|
10077 ms |
|
org.apache.spark.streaming.JavaAPISuite: testFlatMapValues
|
10060 ms |
|
org.apache.spark.streaming.JavaAPISuite: testReduceByWindowWithInverse
|
10232 ms |
|
org.apache.spark.streaming.JavaAPISuite: testMapPartitions
|
10149 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairFilter
|
10104 ms |
|
org.apache.spark.streaming.JavaAPISuite: testRepartitionFewerPartitions
|
10081 ms |
|
org.apache.spark.streaming.JavaAPISuite: testCombineByKey
|
10064 ms |
|
org.apache.spark.streaming.JavaAPISuite: testContextGetOrCreate
|
10131 ms |
|
org.apache.spark.streaming.JavaAPISuite: testWindowWithSlideDuration
|
10058 ms |
|
org.apache.spark.streaming.JavaAPISuite: testQueueStream
|
10065 ms |
|
org.apache.spark.streaming.JavaAPISuite: testCountByValue
|
10067 ms |
|
org.apache.spark.streaming.JavaAPISuite: testMap
|
10146 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairToNormalRDDTransform
|
10069 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairReduceByKey
|
10068 ms |
|
org.apache.spark.streaming.JavaAPISuite: testCount
|
10080 ms |
|
org.apache.spark.streaming.JavaAPISuite: testCheckpointMasterRecovery
|
10078 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairMap
|
10072 ms |
|
org.apache.spark.streaming.JavaAPISuite: testUnion
|
10120 ms |
|
org.apache.spark.streaming.JavaAPISuite: testFlatMap
|
10104 ms |
|
org.apache.spark.streaming.JavaAPISuite: testReduceByKeyAndWindowWithInverse
|
10065 ms |
|
org.apache.spark.streaming.JavaAPISuite: testGlom
|
10069 ms |
|
org.apache.spark.streaming.JavaAPISuite: testJoin
|
10063 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairFlatMap
|
10828 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairToPairFlatMapWithChangingTypes
|
10067 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairMapPartitions
|
10064 ms |
|
org.apache.spark.streaming.JavaAPISuite: testRepartitionMorePartitions
|
10113 ms |
|
org.apache.spark.streaming.JavaAPISuite: testReduceByWindowWithoutInverse
|
10103 ms |
|
org.apache.spark.streaming.JavaAPISuite: testLeftOuterJoin
|
10073 ms |
|
org.apache.spark.streaming.JavaAPISuite: testVariousTransform
|
10071 ms |
|
org.apache.spark.streaming.JavaAPISuite: testTransformWith
|
10158 ms |
|
org.apache.spark.streaming.JavaAPISuite: testVariousTransformWith
|
10076 ms |
|
org.apache.spark.streaming.JavaAPISuite: testTextFileStream
|
10069 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairGroupByKey
|
10075 ms |
|
org.apache.spark.streaming.JavaAPISuite: testCoGroup
|
10188 ms |
|
org.apache.spark.streaming.JavaAPISuite: testInitialization
|
10072 ms |
|
org.apache.spark.streaming.JavaAPISuite: testSocketString
|
10107 ms |
|
org.apache.spark.streaming.JavaAPISuite: testGroupByKeyAndWindow
|
10068 ms |
|
org.apache.spark.streaming.JavaAPISuite: testReduceByKeyAndWindow
|
10064 ms |
|
org.apache.spark.streaming.JavaAPISuite: testForeachRDD
|
11063 ms |
|
org.apache.spark.streaming.JavaAPISuite: testFileStream
|
10073 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairTransform
|
10072 ms |
|
org.apache.spark.streaming.JavaAPISuite: testFilter
|
10082 ms |
|
org.apache.spark.streaming.JavaAPISuite: testPairMap2
|
10070 ms |
|
org.apache.spark.streaming.JavaAPISuite: testMapValues
|
10067 ms |
|
org.apache.spark.streaming.JavaAPISuite: testReduce
|
10110 ms |
|
org.apache.spark.streaming.JavaAPISuite: testUpdateStateByKey
|
10065 ms |
|
org.apache.spark.streaming.JavaAPISuite: testTransform
|
10079 ms |
|
org.apache.spark.streaming.JavaAPISuite: testWindow
|
10758 ms |
|
org.apache.spark.streaming.JavaAPISuite: testCountByValueAndWindow
|
10109 ms |
|
org.apache.spark.streaming.JavaAPISuite: testRawSocketStream
|
10070 ms |
|
org.apache.spark.streaming.JavaAPISuite: testSocketTextStream
|
10071 ms |
|
org.apache.spark.streaming.JavaAPISuite: testUpdateStateByKeyWithInitial
|
10156 ms |
|
org.apache.spark.streaming.JavaAPISuite: testContextState
|
10074 ms |
|
org.apache.spark.streaming.JavaMapWithStateSuite: testBasicFunction
|
10077 ms |
|
org.apache.spark.streaming.JavaReceiverAPISuite: testReceiver
|
10077 ms |
|
org.apache.spark.streaming.MapWithStateSuite: (It is not a test)
|
10058 ms |
|
org.apache.spark.streaming.ReceiverInputDStreamSuite: Without WAL enabled: createBlockRDD creates empty BlockRDD when no block info
|
10037 ms |
|
org.apache.spark.streaming.ReceiverInputDStreamSuite: Without WAL enabled: createBlockRDD creates correct BlockRDD with block info
|
10037 ms |
|
org.apache.spark.streaming.ReceiverInputDStreamSuite: Without WAL enabled: createBlockRDD filters non-existent blocks before creating BlockRDD
|
10060 ms |
|
org.apache.spark.streaming.ReceiverInputDStreamSuite: With WAL enabled: createBlockRDD creates empty WALBackedBlockRDD when no block info
|
10045 ms |
|
org.apache.spark.streaming.ReceiverInputDStreamSuite: With WAL enabled: createBlockRDD creates correct WALBackedBlockRDD with all block info having WAL info
|
10041 ms |
|
org.apache.spark.streaming.ReceiverInputDStreamSuite: With WAL enabled: createBlockRDD creates BlockRDD when some block info dont have WAL info
|
10046 ms |
|
org.apache.spark.streaming.ReceiverSuite: write ahead log - generating and cleaning
|
10034 ms |
|
org.apache.spark.streaming.StreamingContextSuite: from no conf constructor
|
10060 ms |
|
org.apache.spark.streaming.StreamingContextSuite: from no conf + spark home
|
10039 ms |
|
org.apache.spark.streaming.StreamingContextSuite: from no conf + spark home + env
|
10040 ms |
|
org.apache.spark.streaming.StreamingContextSuite: from conf with settings
|
10039 ms |
|
org.apache.spark.streaming.StreamingContextSuite: from existing SparkContext
|
10044 ms |
|
org.apache.spark.streaming.StreamingContextSuite: from existing SparkContext with settings
|
10049 ms |
|
org.apache.spark.streaming.StreamingContextSuite: from checkpoint
|
10039 ms |
|
org.apache.spark.streaming.StreamingContextSuite: checkPoint from conf
|
10045 ms |
|
org.apache.spark.streaming.StreamingContextSuite: start and stop state check
|
10036 ms |
|
org.apache.spark.streaming.StreamingContextSuite: start with non-seriazable DStream checkpoints
|
10037 ms |
|
org.apache.spark.streaming.StreamingContextSuite: start failure should stop internal components
|
10038 ms |
|
org.apache.spark.streaming.StreamingContextSuite: start should set job group and description of streaming jobs correctly
|
10036 ms |
|
org.apache.spark.streaming.StreamingContextSuite: start multiple times
|
10094 ms |
|
org.apache.spark.streaming.StreamingContextSuite: stop multiple times
|
10050 ms |
|
org.apache.spark.streaming.StreamingContextSuite: stop before start
|
10037 ms |
|
org.apache.spark.streaming.StreamingContextSuite: start after stop
|
10060 ms |
|
org.apache.spark.streaming.StreamingContextSuite: stop only streaming context
|
10043 ms |
|
org.apache.spark.streaming.StreamingContextSuite: stop(stopSparkContext=true) after stop(stopSparkContext=false)
|
10041 ms |
|
org.apache.spark.streaming.StreamingContextSuite: stop gracefully
|
10074 ms |
|
org.apache.spark.streaming.StreamingContextSuite: stop gracefully even if a receiver misses StopReceiver
|
10039 ms |
|
org.apache.spark.streaming.StreamingContextSuite: stop slow receiver gracefully
|
10039 ms |
|
org.apache.spark.streaming.StreamingContextSuite: registering and de-registering of streamingSource
|
10036 ms |
|
org.apache.spark.streaming.StreamingContextSuite: awaitTermination
|
10076 ms |
|
org.apache.spark.streaming.StreamingContextSuite: awaitTermination after stop
|
10040 ms |
|
org.apache.spark.streaming.StreamingContextSuite: awaitTermination with error in task
|
10039 ms |
|
org.apache.spark.streaming.StreamingContextSuite: awaitTermination with error in job generation
|
10036 ms |
|
org.apache.spark.streaming.StreamingContextSuite: awaitTerminationOrTimeout
|
10052 ms |
|
org.apache.spark.streaming.StreamingContextSuite: getOrCreate
|
10059 ms |
|
org.apache.spark.streaming.StreamingContextSuite: getActive and getActiveOrCreate
|
10041 ms |
|
org.apache.spark.streaming.StreamingContextSuite: getActiveOrCreate with checkpoint
|
10052 ms |
|
org.apache.spark.streaming.StreamingContextSuite: multiple streaming contexts
|
10057 ms |
|
org.apache.spark.streaming.StreamingContextSuite: DStream and generated RDD creation sites
|
10043 ms |
|
org.apache.spark.streaming.StreamingContextSuite: throw exception on using active or stopped context
|
10056 ms |
|
org.apache.spark.streaming.StreamingContextSuite: queueStream doesn't support checkpointing
|
10312 ms |
|
org.apache.spark.streaming.StreamingContextSuite: Creating an InputDStream but not using it should not crash
|
10036 ms |
|
org.apache.spark.streaming.StreamingListenerSuite: batch info reporting
|
10037 ms |
|
org.apache.spark.streaming.StreamingListenerSuite: receiver info reporting
|
10042 ms |
|
org.apache.spark.streaming.StreamingListenerSuite: output operation reporting
|
10037 ms |
|
org.apache.spark.streaming.StreamingListenerSuite: stop in listener
|
10041 ms |
|
org.apache.spark.streaming.StreamingListenerSuite: onBatchCompleted with successful batch
|
10038 ms |
|
org.apache.spark.streaming.StreamingListenerSuite: onBatchCompleted with failed batch and one failed job
|
10047 ms |
|
org.apache.spark.streaming.StreamingListenerSuite: onBatchCompleted with failed batch and multiple failed jobs
|
10035 ms |
|
org.apache.spark.streaming.UISeleniumSuite: attaching and detaching a Streaming tab
|
10040 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: window - basic window
|
10040 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: window - tumbling window
|
10036 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: window - larger window
|
10040 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: window - non-overlapping window
|
10054 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: window - persistence level
|
10046 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow - basic reduction
|
10036 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow - key already in window and new value added into window
|
10037 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow - new key added into window
|
10047 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow - key removed from window
|
10033 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow - larger slide time
|
10068 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow - big test
|
10038 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow with inverse function - basic reduction
|
10037 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow with inverse function - key already in window and new value added into window
|
10033 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow with inverse function - new key added into window
|
10033 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow with inverse function - key removed from window
|
10034 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow with inverse function - larger slide time
|
10034 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow with inverse function - big test
|
10033 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: reduceByKeyAndWindow with inverse and filter functions - big test
|
10050 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: groupByKeyAndWindow
|
10036 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: countByWindow
|
10034 ms |
|
org.apache.spark.streaming.WindowOperationsSuite: countByValueAndWindow
|
10051 ms |
|
org.apache.spark.streaming.rdd.MapWithStateRDDSuite: (It is not a test)
|
11582 ms |
|
org.apache.spark.streaming.rdd.WriteAheadLogBackedBlockRDDSuite: (It is not a test)
|
10042 ms |
|
org.apache.spark.streaming.scheduler.InputInfoTrackerSuite: (It is not a test)
|
10064 ms |
|
org.apache.spark.streaming.scheduler.RateControllerSuite: RateController - rate controller publishes updates after batches complete
|
10064 ms |
|
org.apache.spark.streaming.scheduler.RateControllerSuite: ReceiverRateController - published rates reach receivers
|
10066 ms |
|
org.apache.spark.streaming.scheduler.ReceiverTrackerSuite: send rate update to receivers
|
10077 ms |
|
org.apache.spark.streaming.scheduler.ReceiverTrackerSuite: should restart receiver after stopping it
|
10061 ms |
|
org.apache.spark.streaming.scheduler.ReceiverTrackerSuite: SPARK-11063: TaskSetManager should use Receiver RDD's preferredLocations
|
10061 ms |
|
org.apache.spark.streaming.ui.StreamingJobProgressListenerSuite: onBatchSubmitted, onBatchStarted, onBatchCompleted, onReceiverStarted, onReceiverError, onReceiverStopped
|
10071 ms |
|
org.apache.spark.streaming.ui.StreamingJobProgressListenerSuite: Remove the old completed batches when exceeding the limit
|
10046 ms |
|
org.apache.spark.streaming.ui.StreamingJobProgressListenerSuite: out-of-order onJobStart and onBatchXXX
|
10050 ms |
|
org.apache.spark.streaming.ui.StreamingJobProgressListenerSuite: detect memory leak
|
10040 ms |
|
org.apache.spark.deploy.yarn.YarnClusterSuite: run Spark in yarn-client mode
|
120066 ms |
|
org.apache.spark.deploy.yarn.YarnClusterSuite: run Spark in yarn-cluster mode
|
53034 ms |
|
org.apache.spark.deploy.yarn.YarnClusterSuite: run Python application in yarn-client mode
|
120081 ms |
|
org.apache.spark.deploy.yarn.YarnClusterSuite: run Python application in yarn-cluster mode
|
52035 ms |
|
org.apache.spark.deploy.yarn.YarnClusterSuite: user class path first in client mode
|
120062 ms |
|
org.apache.spark.deploy.yarn.YarnClusterSuite: user class path first in cluster mode
|
53037 ms |
|
org.apache.spark.deploy.yarn.YarnClusterSuite: monitor app using launcher library
|
30081 ms |
|
org.apache.spark.deploy.yarn.YarnShuffleIntegrationSuite: external shuffle service
|
53096 ms |
|
org.apache.spark.scheduler.cluster.ExtensionServiceIntegrationSuite: (It is not a test)
|
11603 ms |
|