[MINOR] improvement(server) Add context to rpc audit log to output necessary context #2088
3 errors, 1 fail, 1 skipped, 1 014 pass in 5h 30m 16s
Annotations
Check failure on line 0 in org.apache.uniffle.test.RepartitionWithMemoryRssTest
github-actions / Test Results
1 out of 10 runs with error: resultCompareTest (org.apache.uniffle.test.RepartitionWithMemoryRssTest)
artifacts/integration-reports-spark3.5-scala2.13/integration-test/spark-common/target/surefire-reports/TEST-org.apache.uniffle.test.RepartitionWithMemoryRssTest.xml [took 7s]
Raw output
Job aborted due to stage failure: Task 1 in stage 1.0 failed 1 times, most recent failure: Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
org.apache.spark.SparkException:
Job aborted due to stage failure: Task 1 in stage 1.0 failed 1 times, most recent failure: Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2856)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2792)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2791)
at scala.collection.immutable.List.foreach(List.scala:333)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2791)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1247)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1247)
at scala.Option.foreach(Option.scala:437)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1247)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:3060)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2994)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2983)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:989)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2398)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2419)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2438)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2463)
at org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1049)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:410)
at org.apache.spark.rdd.RDD.collect(RDD.scala:1048)
at org.apache.spark.RangePartitioner$.sketch(Partitioner.scala:320)
at org.apache.spark.RangePartitioner.<init>(Partitioner.scala:187)
at org.apache.spark.RangePartitioner.<init>(Partitioner.scala:167)
at org.apache.spark.rdd.OrderedRDDFunctions.$anonfun$sortByKey$1(OrderedRDDFunctions.scala:64)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:410)
at org.apache.spark.rdd.OrderedRDDFunctions.sortByKey(OrderedRDDFunctions.scala:63)
at org.apache.spark.api.java.JavaPairRDD.sortByKey(JavaPairRDD.scala:927)
at org.apache.spark.api.java.JavaPairRDD.sortByKey(JavaPairRDD.scala:897)
at org.apache.spark.api.java.JavaPairRDD.sortByKey(JavaPairRDD.scala:887)
at org.apache.uniffle.test.RepartitionTest.repartitionApp(RepartitionTest.java:99)
at org.apache.uniffle.test.RepartitionTest.runTest(RepartitionTest.java:49)
at org.apache.uniffle.test.SparkIntegrationTestBase.runSparkApp(SparkIntegrationTestBase.java:102)
at org.apache.uniffle.test.SparkIntegrationTestBase.run(SparkIntegrationTestBase.java:62)
at org.apache.uniffle.test.RepartitionTest.resultCompareTest(RepartitionTest.java:44)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725)
at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:107)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:150)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:124)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
Caused by: org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:00.154] [main] [INFO] MiniDFSCluster.<init> - starting cluster: numNameNodes=1, numDataNodes=1
Formatting using clusterid: testClusterID
[2024-09-20 04:16:00.154] [main] [INFO] FSEditLog.newInstance - Edit logging is async:true
[2024-09-20 04:16:00.155] [main] [INFO] FSNamesystem.<init> - KeyProvider: null
[2024-09-20 04:16:00.155] [main] [INFO] FSNamesystem.<init> - fsLock is fair: true
[2024-09-20 04:16:00.155] [main] [INFO] FSNamesystem.<init> - Detailed lock hold time metrics enabled: false
[2024-09-20 04:16:00.155] [main] [INFO] DatanodeManager.<init> - dfs.block.invalidate.limit=1000
[2024-09-20 04:16:00.155] [main] [INFO] DatanodeManager.<init> - dfs.namenode.datanode.registration.ip-hostname-check=true
[2024-09-20 04:16:00.155] [main] [INFO] BlockManager.printBlockDeletionTime - dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
[2024-09-20 04:16:00.156] [main] [INFO] BlockManager.printBlockDeletionTime - The block deletion will start around 2024 Sep 20 04:16:00
[2024-09-20 04:16:00.156] [main] [INFO] GSet.computeCapacity - Computing capacity for map BlocksMap
[2024-09-20 04:16:00.156] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:16:00.156] [main] [INFO] GSet.computeCapacity - 2.0% max memory 4.4 GB = 91.0 MB
[2024-09-20 04:16:00.156] [main] [INFO] GSet.computeCapacity - capacity = 2^24 = 16777216 entries
[2024-09-20 04:16:00.157] [main] [INFO] BlockManager.createBlockTokenSecretManager - dfs.block.access.token.enable=false
[2024-09-20 04:16:00.157] [main] [INFO] BlockManager.<init> - defaultReplication = 1
[2024-09-20 04:16:00.157] [main] [INFO] BlockManager.<init> - maxReplication = 512
[2024-09-20 04:16:00.158] [main] [INFO] BlockManager.<init> - minReplication = 1
[2024-09-20 04:16:00.158] [main] [INFO] BlockManager.<init> - maxReplicationStreams = 2
[2024-09-20 04:16:00.158] [main] [INFO] BlockManager.<init> - replicationRecheckInterval = 3000
[2024-09-20 04:16:00.158] [main] [INFO] BlockManager.<init> - encryptDataTransfer = false
[2024-09-20 04:16:00.158] [main] [INFO] BlockManager.<init> - maxNumBlocksToLog = 1000
[2024-09-20 04:16:00.158] [main] [INFO] FSNamesystem.<init> - fsOwner = runner (auth:SIMPLE)
[2024-09-20 04:16:00.158] [main] [INFO] FSNamesystem.<init> - supergroup = supergroup
[2024-09-20 04:16:00.158] [main] [INFO] FSNamesystem.<init> - isPermissionEnabled = true
[2024-09-20 04:16:00.158] [main] [INFO] FSNamesystem.<init> - HA Enabled: false
[2024-09-20 04:16:00.158] [main] [INFO] FSNamesystem.<init> - Append Enabled: true
[2024-09-20 04:16:00.158] [main] [INFO] GSet.computeCapacity - Computing capacity for map INodeMap
[2024-09-20 04:16:00.159] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:16:00.159] [main] [INFO] GSet.computeCapacity - 1.0% max memory 4.4 GB = 45.5 MB
[2024-09-20 04:16:00.159] [main] [INFO] GSet.computeCapacity - capacity = 2^23 = 8388608 entries
[2024-09-20 04:16:00.160] [main] [INFO] FSDirectory.<init> - ACLs enabled? false
[2024-09-20 04:16:00.160] [main] [INFO] FSDirectory.<init> - XAttrs enabled? true
[2024-09-20 04:16:00.161] [main] [INFO] NameNode.<init> - Caching file names occurring more than 10 times
[2024-09-20 04:16:00.161] [main] [INFO] GSet.computeCapacity - Computing capacity for map cachedBlocks
[2024-09-20 04:16:00.161] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:16:00.161] [main] [INFO] GSet.computeCapacity - 0.25% max memory 4.4 GB = 11.4 MB
[2024-09-20 04:16:00.161] [main] [INFO] GSet.computeCapacity - capacity = 2^21 = 2097152 entries
[2024-09-20 04:16:00.161] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.threshold-pct = 0.9990000128746033
[2024-09-20 04:16:00.161] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.min.datanodes = 0
[2024-09-20 04:16:00.161] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.extension = 0
[2024-09-20 04:16:00.162] [main] [INFO] TopMetrics.logConf - NNTop conf: dfs.namenode.top.window.num.buckets = 10
[2024-09-20 04:16:00.162] [main] [INFO] TopMetrics.logConf - NNTop conf: dfs.namenode.top.num.users = 10
[2024-09-20 04:16:00.162] [main] [INFO] TopMetrics.logConf - NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25
[2024-09-20 04:16:00.162] [main] [INFO] FSNamesystem.initRetryCache - Retry cache on namenode is enabled
[2024-09-20 04:16:00.162] [main] [INFO] FSNamesystem.initRetryCache - Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
[2024-09-20 04:16:00.162] [main] [INFO] GSet.computeCapacity - Computing capacity for map NameNodeRetryCache
[2024-09-20 04:16:00.162] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:16:00.162] [main] [INFO] GSet.computeCapacity - 0.029999999329447746% max memory 4.4 GB = 1.4 MB
[2024-09-20 04:16:00.162] [main] [INFO] GSet.computeCapacity - capacity = 2^17 = 131072 entries
[2024-09-20 04:16:00.163] [main] [INFO] FSImage.format - Allocated new BlockPoolId: BP-1092480937-127.0.0.1-1726805760163
[2024-09-20 04:16:00.165] [main] [INFO] Storage.format - Storage directory /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name1 has been successfully formatted.
[2024-09-20 04:16:00.166] [main] [INFO] Storage.format - Storage directory /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name2 has been successfully formatted.
[2024-09-20 04:16:00.166] [FSImageSaver for /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name1 of type IMAGE_AND_EDITS] [INFO] FSImageFormatProtobuf.save - Saving image file /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name1/current/fsimage.ckpt_0000000000000000000 using no compression
[2024-09-20 04:16:00.166] [FSImageSaver for /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name2 of type IMAGE_AND_EDITS] [INFO] FSImageFormatProtobuf.save - Saving image file /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name2/current/fsimage.ckpt_0000000000000000000 using no compression
[2024-09-20 04:16:00.178] [FSImageSaver for /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name2 of type IMAGE_AND_EDITS] [INFO] FSImageFormatProtobuf.save - Image file /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name2/current/fsimage.ckpt_0000000000000000000 of size 323 bytes saved in 0 seconds.
[2024-09-20 04:16:00.178] [FSImageSaver for /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name1 of type IMAGE_AND_EDITS] [INFO] FSImageFormatProtobuf.save - Image file /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit1780316645609527324/name1/current/fsimage.ckpt_0000000000000000000 of size 323 bytes saved in 0 seconds.
[2024-09-20 04:16:00.179] [main] [INFO] NNStorageRetentionManager.getImageTxIdToRetain - Going to retain 1 images with txid >= 0
[2024-09-20 04:16:00.180] [main] [INFO] NameNode.createNameNode - createNameNode []
[2024-09-20 04:16:00.181] [main] [WARN] MetricsConfig.loadFirst - Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[2024-09-20 04:16:00.182] [main] [INFO] MetricsSystemImpl.startTimer - Scheduled Metric snapshot period at 10 second(s).
[2024-09-20 04:16:00.182] [main] [INFO] MetricsSystemImpl.start - NameNode metrics system started
[2024-09-20 04:16:00.182] [main] [INFO] NameNode.setClientNamenodeAddress - fs.defaultFS is hdfs://127.0.0.1:0
[2024-09-20 04:16:00.187] [org.apache.hadoop.util.JvmPauseMonitor$Monitor@2841c0ca] [INFO] JvmPauseMonitor.run - Starting JVM pause monitor
[2024-09-20 04:16:00.187] [main] [INFO] DFSUtil.httpServerTemplateForNNAndJN - Starting Web-server for hdfs at: http://localhost:0
[2024-09-20 04:16:00.188] [main] [INFO] AuthenticationFilter.constructSecretProvider - Unable to initialize FileSignerSecretProvider, falling back to use random secrets.
[2024-09-20 04:16:00.188] [main] [WARN] HttpRequestLog.getRequestLog - Jetty request log can only be enabled using Log4j
[2024-09-20 04:16:00.189] [main] [INFO] HttpServer2.addGlobalFilter - Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
[2024-09-20 04:16:00.189] [main] [INFO] HttpServer2.addFilter - Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context hdfs
[2024-09-20 04:16:00.189] [main] [INFO] HttpServer2.addFilter - Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
[2024-09-20 04:16:00.189] [main] [INFO] HttpServer2.initWebHdfs - Added filter 'org.apache.hadoop.hdfs.web.AuthFilter' (class=org.apache.hadoop.hdfs.web.AuthFilter)
[2024-09-20 04:16:00.189] [main] [INFO] HttpServer2.addJerseyResourcePackage - addJerseyResourcePackage: packageName=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources, pathSpec=/webhdfs/v1/*
[2024-09-20 04:16:00.190] [main] [INFO] HttpServer2.openListeners - Jetty bound to port 44089
[2024-09-20 04:16:00.190] [main] [INFO] log.info - jetty-6.1.26
[2024-09-20 04:16:00.194] [main] [INFO] log.info - Extract jar:file:/home/runner/.m2/repository/org/apache/hadoop/hadoop-hdfs/2.8.5/hadoop-hdfs-2.8.5-tests.jar!/webapps/hdfs to /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/Jetty_localhost_44089_hdfs____.suak6d/webapp
[2024-09-20 04:16:00.258] [main] [INFO] log.info - Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:44089
[2024-09-20 04:16:00.258] [main] [INFO] FSEditLog.newInstance - Edit logging is async:true
[2024-09-20 04:16:00.258] [main] [INFO] FSNamesystem.<init> - KeyProvider: null
[2024-09-20 04:16:00.258] [main] [INFO] FSNamesystem.<init> - fsLock is fair: true
[2024-09-20 04:16:00.258] [main] [INFO] FSNamesystem.<init> - Detailed lock hold time metrics enabled: false
[2024-09-20 04:16:00.259] [main] [INFO] DatanodeManager.<init> - dfs.block.invalidate.limit=1000
[2024-09-20 04:16:00.259] [main] [INFO] DatanodeManager.<init> - dfs.namenode.datanode.registration.ip-hostname-check=true
[2024-09-20 04:16:00.259] [main] [INFO] BlockManager.printBlockDeletionTime - dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
[2024-09-20 04:16:00.259] [main] [INFO] BlockManager.printBlockDeletionTime - The block deletion will start around 2024 Sep 20 04:16:00
[2024-09-20 04:16:00.259] [main] [INFO] GSet.computeCapacity - Computing capacity for map BlocksMap
[2024-09-20 04:16:00.259] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:16:00.260] [main] [INFO] GSet.computeCapacity - 2.0% max memory 4.4 GB = 91.0 MB
[2024-09-20 04:16:00.260] [main] [INFO] GSet.computeCapacity - capacity = 2^24 = 16777216 entries
[2024-09-20 04:16:00.299] [main] [INFO] BlockManager.createBlockTokenSecretManager - dfs.block.access.token.enable=false
[2024-09-20 04:16:00.299] [main] [INFO] BlockManager.<init> - defaultReplication = 1
[2024-09-20 04:16:00.300] [main] [INFO] BlockManager.<init> - maxReplication = 512
[2024-09-20 04:16:00.300] [main] [INFO] BlockManager.<init> - minReplication = 1
[2024-09-20 04:16:00.300] [main] [INFO] BlockManager.<init> - maxReplicationStreams = 2
[2024-09-20 04:16:00.300] [main] [INFO] BlockManager.<init> - replicationRecheckInterval = 3000
[2024-09-20 04:16:00.300] [main] [INFO] BlockManager.<init> - encryptDataTransfer = false
[2024-09-20 04:16:00.300] [main] [INFO] BlockManager.<init> - maxNumBlocksToLog = 1000
[2024-09-20 04:16:00.300] [main] [INFO] FSNamesystem.<init> - fsOwner = runner (auth:SIMPLE)
[2024-09-20 04:16:00.300] [main] [INFO] FSNamesystem.<init> - supergroup = supergroup
[2024-09-20 04:16:00.300] [main] [INFO] FSNamesystem.<init> - isPermissionEnabled = true
[2024-09-20 04:16:00.300] [main] [INFO] FSNamesystem.<init> - HA Enabled: false
[2024-09-20 04:16:00.300] [main] [INFO] FSNamesystem.<init> - Append Enabled: true
[2024-09-20 04:16:00.301] [main] [INFO] GSet.computeCapacity - Computing capacity for map INodeMap
[2024-09-20 04:16:00.301] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:16:00.301] [main] [INFO] GSet.computeCapacity - 1.0% max memory 4.4 GB = 45.5 MB
[2024-09-20 04:16:00.301] [main] [INFO] GSet.computeCapacity - capacity = 2^23 = 8388608 entries
[2024-09-20 04:16:00.303] [main] [INFO] FSDirectory.<init> - ACLs enabled? false
[2024-09-20 04:16:00.303] [main] [INFO] FSDirectory.<init> - XAttrs enabled? true
[2024-09-20 04:16:00.304] [main] [INFO] NameNode.<init> - Caching file names occurring more than 10 times
[2024-09-20 04:16:00.304] [main] [INFO] GSet.computeCapacity - Computing capacity for map cachedBlocks
[2024-09-20 04:16:00.304] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:16:00.304] [main] [INFO] GSet.computeCapacity - 0.25% max memory 4.4 GB = 11.4 MB
[2024-09-20 04:16:00.304] [main] [INFO] GSet.computeCapacity - capacity = 2^21 = 2097152 entries
[2024-09-20 04:16:00.305] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.threshold-pct = 0.9990000128746033
[2024-09-20 04:16:00.305] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.min.datanodes = 0
[2024-09-20 04:16:00.305] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.extension = 0
[2024-09-20 04:16:00.305] [main] [INFO] TopMetrics.logConf - NNTop conf: dfs.namenode.top.window.num.buckets = 10
[2024-09-20 04:16:00.305] [main] [INFO] TopMetrics.logConf - NNTop conf: dfs.namenode.top.num.users = 10
[2024-09-20 04:16:00.305] [main] [INFO] TopMetrics.logConf - NNTop conf: …in stage 1.0 (TID 7)] [INFO] RssShuffleWriter.<init> - RssShuffle start write taskAttemptId[12] data with RssHandle[appId local-1726805768964_1726805768940, shuffleId 1].
[2024-09-20 04:16:10.342] [Executor task launch worker for task 0.0 in stage 1.0 (TID 4)] [INFO] RssShuffleManager.getReader - Get taskId cost 0 ms, and request expected blockIds from 4 tasks for shuffleId[0], partitionId[0, 1]
[2024-09-20 04:16:10.342] [Executor task launch worker for task 1.0 in stage 1.0 (TID 5)] [INFO] RssShuffleManager.getReader - Get taskId cost 0 ms, and request expected blockIds from 4 tasks for shuffleId[0], partitionId[1, 2]
[2024-09-20 04:16:10.342] [Grpc-2] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=SUCCESS from=/10.1.0.21:41108 executionTimeUs=138 appId=local-1726805768964_1726805768940 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=59} context{bitmap[0].<size,byte>=<35,310>, partitionBlockCount=7}
[2024-09-20 04:16:10.342] [Grpc-3] [ERROR] ShuffleServerGrpcService.getShuffleResultForMultiPart - Error happened when get shuffle result for appId[local-1726805768964_1726805768940], shuffleId[0], partitions[1]
java.lang.ArrayIndexOutOfBoundsException: 0
at org.roaringbitmap.longlong.Roaring64NavigableMap.ensureOne(Roaring64NavigableMap.java:676)
at org.roaringbitmap.longlong.Roaring64NavigableMap.ensureCumulatives(Roaring64NavigableMap.java:567)
at org.roaringbitmap.longlong.Roaring64NavigableMap.getLongCardinality(Roaring64NavigableMap.java:278)
at org.apache.uniffle.server.ShuffleTaskManager.lambda$getFinishedBlockIds$9(ShuffleTaskManager.java:657)
at java.util.Optional.ifPresent(Optional.java:159)
at org.apache.uniffle.server.ShuffleTaskManager.getFinishedBlockIds(ShuffleTaskManager.java:652)
at org.apache.uniffle.server.ShuffleServerGrpcService.getShuffleResultForMultiPart(ShuffleServerGrpcService.java:985)
at org.apache.uniffle.proto.ShuffleServerGrpc$MethodHandlers.invoke(ShuffleServerGrpc.java:1180)
at io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:182)
at io.grpc.PartialForwardingServerCallListener.onHalfClose(PartialForwardingServerCallListener.java:35)
at io.grpc.ForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:23)
at io.grpc.ForwardingServerCallListener$SimpleForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:40)
at org.apache.uniffle.common.rpc.ClientContextServerInterceptor$1.onHalfClose(ClientContextServerInterceptor.java:63)
at io.grpc.PartialForwardingServerCallListener.onHalfClose(PartialForwardingServerCallListener.java:35)
at io.grpc.ForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:23)
at io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:356)
at io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:861)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:10.343] [Executor task launch worker for task 0.0 in stage 1.0 (TID 4)] [INFO] RssShuffleManager.getReaderImpl - Get shuffle blockId cost 1 ms, and get 7 blockIds for shuffleId[0], startPartition[0], endPartition[1]
[2024-09-20 04:16:10.343] [Grpc-3] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=INTERNAL_ERROR from=/10.1.0.21:41108 executionTimeUs=389 appId=local-1726805768964_1726805768940 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=0}
[2024-09-20 04:16:10.343] [Executor task launch worker for task 0.0 in stage 1.0 (TID 4)] [INFO] RssShuffleManager.getReaderImpl - Shuffle reader using remote storage Empty Remote Storage
[2024-09-20 04:16:10.343] [Executor task launch worker for task 2.0 in stage 1.0 (TID 6)] [INFO] RssShuffleManager.getReader - Get taskId cost 0 ms, and request expected blockIds from 4 tasks for shuffleId[0], partitionId[2, 3]
[2024-09-20 04:16:10.346] [Executor task launch worker for task 3.0 in stage 1.0 (TID 7)] [INFO] RssShuffleManager.getReader - Get taskId cost 0 ms, and request expected blockIds from 4 tasks for shuffleId[0], partitionId[3, 4]
[2024-09-20 04:16:10.350] [Executor task launch worker for task 1.0 in stage 1.0 (TID 5)] [ERROR] ShuffleServerGrpcClient.getShuffleResultForMultiPart - Can't get shuffle result from 10.1.0.21:20045 for [appId=local-1726805768964_1726805768940, shuffleId=0, errorMsg:0
[2024-09-20 04:16:10.350] [Grpc-9] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=SUCCESS from=/10.1.0.21:41108 executionTimeUs=82 appId=local-1726805768964_1726805768940 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=59} context{bitmap[0].<size,byte>=<35,310>, partitionBlockCount=7}
[2024-09-20 04:16:10.350] [Grpc-0] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=SUCCESS from=/10.1.0.21:41108 executionTimeUs=68 appId=local-1726805768964_1726805768940 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=59} context{bitmap[0].<size,byte>=<35,310>, partitionBlockCount=7}
[2024-09-20 04:16:10.350] [Executor task launch worker for task 1.0 in stage 1.0 (TID 5)] [WARN] ShuffleWriteClientImpl.getShuffleResultForMultiPart - Get shuffle result is failed from ShuffleServerInfo{host[10.1.0.21], grpc port[20045]} for appId[local-1726805768964_1726805768940], shuffleId[0], requestPartitions[1]
org.apache.uniffle.common.exception.RssFetchFailedException: Can't get shuffle result from 10.1.0.21:20045 for [appId=local-1726805768964_1726805768940, shuffleId=0, errorMsg:0
at org.apache.uniffle.client.impl.grpc.ShuffleServerGrpcClient.getShuffleResultForMultiPart(ShuffleServerGrpcClient.java:913)
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:860)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:10.350] [Executor task launch worker for task 3.0 in stage 1.0 (TID 7)] [INFO] RssShuffleManager.getReaderImpl - Get shuffle blockId cost 4 ms, and get 7 blockIds for shuffleId[0], startPartition[3], endPartition[4]
[2024-09-20 04:16:10.350] [Executor task launch worker for task 1.0 in stage 1.0 (TID 5)] [ERROR] ShuffleWriteClientImpl.getShuffleResultForMultiPart - Failed to meet replica requirement: PartitionDataReplicaRequirementTracking{shuffleId=0, inventory={0={0=[ShuffleServerInfo{host[10.1.0.21], grpc port[20045]}]}, 1={0=[ShuffleServerInfo{host[10.1.0.21], grpc port[20045]}]}, 2={0=[ShuffleServerInfo{host[10.1.0.21], grpc port[20045]}]}, 3={0=[ShuffleServerInfo{host[10.1.0.21], grpc port[20045]}]}, 4={0=[ShuffleServerInfo{host[10.1.0.21], grpc port[20045]}]}}, succeedList={}}
[2024-09-20 04:16:10.350] [Executor task launch worker for task 3.0 in stage 1.0 (TID 7)] [INFO] RssShuffleManager.getReaderImpl - Shuffle reader using remote storage Empty Remote Storage
[2024-09-20 04:16:10.350] [Executor task launch worker for task 1.0 in stage 1.0 (TID 5)] [INFO] RssShuffleManager.markFailedTask - Mark the task: 5_0 failed.
[2024-09-20 04:16:10.350] [Executor task launch worker for task 1.0 in stage 1.0 (TID 5)] [ERROR] Executor.logError - Exception in task 1.0 in stage 1.0 (TID 5)
org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:10.351] [dispatcher-event-loop-1] [INFO] TaskSetManager.logInfo - Starting task 4.0 in stage 1.0 (TID 8) (fv-az1981-56, executor driver, partition 4, ANY, 7849 bytes)
[2024-09-20 04:16:10.352] [Executor task launch worker for task 2.0 in stage 1.0 (TID 6)] [INFO] RssShuffleManager.getReaderImpl - Get shuffle blockId cost 9 ms, and get 7 blockIds for shuffleId[0], startPartition[2], endPartition[3]
[2024-09-20 04:16:10.352] [task-result-getter-0] [WARN] TaskSetManager.logWarning - Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:10.352] [Executor task launch worker for task 2.0 in stage 1.0 (TID 6)] [INFO] RssShuffleManager.getReaderImpl - Shuffle reader using remote storage Empty Remote Storage
[2024-09-20 04:16:10.352] [task-result-getter-0] [ERROR] TaskSetManager.logError - Task 1 in stage 1.0 failed 1 times; aborting job
[2024-09-20 04:16:10.352] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Cancelling stage 1
[2024-09-20 04:16:10.352] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Killing all running tasks in stage 1: Stage cancelled: Job aborted due to stage failure: Task 1 in stage 1.0 failed 1 times, most recent failure: Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
[2024-09-20 04:16:10.353] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Stage 1 was cancelled
[2024-09-20 04:16:10.353] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - ShuffleMapStage 1 (repartition at RepartitionTest.java:97) failed in 0.020 s due to Job aborted due to stage failure: Task 1 in stage 1.0 failed 1 times, most recent failure: Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
[2024-09-20 04:16:10.353] [main] [INFO] DAGScheduler.logInfo - Job 0 failed: sortByKey at RepartitionTest.java:99, took 1.316714 s
[2024-09-20 04:16:10.357] [dispatcher-event-loop-1] [INFO] Executor.logInfo - Executor is trying to kill task 0.0 in stage 1.0 (TID 4), reason: Stage cancelled: Job aborted due to stage failure: Task 1 in stage 1.0 failed 1 times, most recent failure: Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
[2024-09-20 04:16:10.357] [dispatcher-event-loop-1] [INFO] Executor.logInfo - Executor is trying to kill task 2.0 in stage 1.0 (TID 6), reason: Stage cancelled: Job aborted due to stage failure: Task 1 in stage 1.0 failed 1 times, most recent failure: Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
[2024-09-20 04:16:10.357] [dispatcher-event-loop-1] [INFO] Executor.logInfo - Executor is trying to kill task 3.0 in stage 1.0 (TID 7), reason: Stage cancelled: Job aborted due to stage failure: Task 1 in stage 1.0 failed 1 times, most recent failure: Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
[2024-09-20 04:16:10.357] [dispatcher-event-loop-1] [INFO] Executor.logInfo - Executor is trying to kill task 4.0 in stage 1.0 (TID 8), reason: Stage cancelled: Job aborted due to stage failure: Task 1 in stage 1.0 failed 1 times, most recent failure: Lost task 1.0 in stage 1.0 (TID 5) (fv-az1981-56 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805768964_1726805768940], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:587)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:601)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at scala.collection.Iterator$$anon$9.hasNext(Iterator.scala:576)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
[2024-09-20 04:16:10.358] [Executor task launch worker for task 4.0 in stage 1.0 (TID 8)] [INFO] Executor.logInfo - Running task 4.0 in stage 1.0 (TID 8)
Check failure on line 0 in org.apache.uniffle.test.RepartitionWithMemoryRssTest
github-actions / Test Results
5 out of 10 runs with error: testMemoryRelease (org.apache.uniffle.test.RepartitionWithMemoryRssTest)
artifacts/integration-reports-spark3.3/integration-test/spark-common/target/surefire-reports/TEST-org.apache.uniffle.test.RepartitionWithMemoryRssTest.xml [took 1m 17s]
artifacts/integration-reports-spark3.4/integration-test/spark-common/target/surefire-reports/TEST-org.apache.uniffle.test.RepartitionWithMemoryRssTest.xml [took 37s]
artifacts/integration-reports-spark3.5-scala2.13/integration-test/spark-common/target/surefire-reports/TEST-org.apache.uniffle.test.RepartitionWithMemoryRssTest.xml [took 35s]
artifacts/integration-reports-spark3.5/integration-test/spark-common/target/surefire-reports/TEST-org.apache.uniffle.test.RepartitionWithMemoryRssTest.xml [took 37s]
artifacts/integration-reports-spark3/integration-test/spark-common/target/surefire-reports/TEST-org.apache.uniffle.test.RepartitionWithMemoryRssTest.xml [took 36s]
Raw output
Job aborted due to stage failure: Task 0 in stage 1.0 failed 1 times, most recent failure: Lost task 0.0 in stage 1.0 (TID 5) (fv-az1775-366 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805795493_1726805795466], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:364)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:328)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:101)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
at org.apache.spark.scheduler.Task.run(Task.scala:139)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
org.apache.spark.SparkException:
Job aborted due to stage failure: Task 0 in stage 1.0 failed 1 times, most recent failure: Lost task 0.0 in stage 1.0 (TID 5) (fv-az1775-366 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805795493_1726805795466], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:364)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:328)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:101)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
at org.apache.spark.scheduler.Task.run(Task.scala:139)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2785)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2721)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2720)
at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2720)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1206)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1206)
at scala.Option.foreach(Option.scala:407)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1206)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2984)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2923)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2912)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:971)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2263)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2284)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2303)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2328)
at org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1019)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:405)
at org.apache.spark.rdd.RDD.collect(RDD.scala:1018)
at org.apache.spark.RangePartitioner$.sketch(Partitioner.scala:320)
at org.apache.spark.RangePartitioner.<init>(Partitioner.scala:187)
at org.apache.spark.RangePartitioner.<init>(Partitioner.scala:167)
at org.apache.spark.rdd.OrderedRDDFunctions.$anonfun$sortByKey$1(OrderedRDDFunctions.scala:64)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:405)
at org.apache.spark.rdd.OrderedRDDFunctions.sortByKey(OrderedRDDFunctions.scala:63)
at org.apache.spark.api.java.JavaPairRDD.sortByKey(JavaPairRDD.scala:927)
at org.apache.spark.api.java.JavaPairRDD.sortByKey(JavaPairRDD.scala:897)
at org.apache.spark.api.java.JavaPairRDD.sortByKey(JavaPairRDD.scala:887)
at org.apache.uniffle.test.RepartitionTest.repartitionApp(RepartitionTest.java:99)
at org.apache.uniffle.test.RepartitionTest.runTest(RepartitionTest.java:49)
at org.apache.uniffle.test.SparkIntegrationTestBase.runSparkApp(SparkIntegrationTestBase.java:102)
at org.apache.uniffle.test.RepartitionWithMemoryRssTest.testMemoryRelease(RepartitionWithMemoryRssTest.java:72)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725)
at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:107)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:150)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:124)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
Caused by: org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805795493_1726805795466], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:364)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:328)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:101)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
at org.apache.spark.scheduler.Task.run(Task.scala:139)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:23.868] [main] [INFO] RepartitionTest.generateTextFile - Create file:/home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit2997157596037892570/wordcount.txt
[2024-09-20 04:16:24.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:25.129] [ApplicationManager-0] [INFO] ApplicationManager.statusCheck - Start to check status for 2 applications.
[2024-09-20 04:16:25.129] [ApplicationManager-0] [INFO] ApplicationManager.statusCheck - Remove expired application : local-1726805774016_1726805773994.
[2024-09-20 04:16:25.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:26.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:27.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:27.629] [ApplicationManager-0] [INFO] ApplicationManager.statusCheck - Start to check status for 1 applications.
[2024-09-20 04:16:28.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:29.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:30.129] [ApplicationManager-0] [INFO] ApplicationManager.statusCheck - Start to check status for 1 applications.
[2024-09-20 04:16:30.129] [ApplicationManager-0] [INFO] ApplicationManager.statusCheck - Remove expired application : local-1726805779094_1726805779065.
[2024-09-20 04:16:30.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:31.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:16:32.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at or…TimeUs=21 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireSize=1774219, partitionIdsSize=1} return{requireBufferId=487}
[2024-09-20 04:17:00.798] [Grpc-9] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=sendShuffleData statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=44 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireBufferId=487, timestamp=1726805820796, stageAttemptNumber=0, shuffleDataSize=1}
[2024-09-20 04:17:01.209] [DynamicClientConfService-0] [WARN] DynamicClientConfService.refreshClientConf - Error when update client conf with hdfs://localhost:35533/test/client_conf.
java.io.IOException: Filesystem closed
at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:466)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1628)
at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1614)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:900)
at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:114)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:964)
at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:961)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:961)
at org.apache.uniffle.coordinator.conf.DynamicClientConfService.refreshClientConf(DynamicClientConfService.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:17:01.353] [Grpc-6] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=requireBuffer statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=33 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireSize=1773679, partitionIdsSize=1} return{requireBufferId=488}
[2024-09-20 04:17:01.355] [Grpc-1] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=sendShuffleData statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=67 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireBufferId=488, timestamp=1726805821353, stageAttemptNumber=0, shuffleDataSize=1}
[2024-09-20 04:17:01.360] [Grpc-3] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=requireBuffer statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=25 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireSize=1774599, partitionIdsSize=1} return{requireBufferId=489}
[2024-09-20 04:17:01.361] [Grpc-8] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=sendShuffleData statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=52 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireBufferId=489, timestamp=1726805821360, stageAttemptNumber=0, shuffleDataSize=1}
[2024-09-20 04:17:01.367] [Grpc-9] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=requireBuffer statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=26 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireSize=1774851, partitionIdsSize=1} return{requireBufferId=490}
[2024-09-20 04:17:01.369] [Grpc-6] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=sendShuffleData statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=54 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireBufferId=490, timestamp=1726805821367, stageAttemptNumber=0, shuffleDataSize=1}
[2024-09-20 04:17:01.374] [Grpc-1] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=requireBuffer statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=21 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireSize=1774364, partitionIdsSize=1} return{requireBufferId=491}
[2024-09-20 04:17:01.375] [Grpc-3] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=sendShuffleData statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=54 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireBufferId=491, timestamp=1726805821374, stageAttemptNumber=0, shuffleDataSize=1}
[2024-09-20 04:17:01.381] [Grpc-4] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=requireBuffer statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=19 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireSize=1774413, partitionIdsSize=1} return{requireBufferId=492}
[2024-09-20 04:17:01.382] [Grpc-0] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=sendShuffleData statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=42 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireBufferId=492, timestamp=1726805821381, stageAttemptNumber=0, shuffleDataSize=1}
[2024-09-20 04:17:01.450] [Executor task launch worker for task 4.0 in stage 0.0 (TID 4)] [INFO] WriteBufferManager.clear - Flush total buffer for shuffleId[0] with allocated[5767168], dataSize[2562384], memoryUsed[2621440], number of blocks[5], flush ratio[1.0]
[2024-09-20 04:17:01.450] [Grpc-5] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=requireBuffer statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=31 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireSize=1084989, partitionIdsSize=5} return{requireBufferId=493}
[2024-09-20 04:17:01.451] [Grpc-2] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=sendShuffleData statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=63 appId=local-1726805795493_1726805795466 shuffleId=0 args{requireBufferId=493, timestamp=1726805821450, stageAttemptNumber=0, shuffleDataSize=5}
[2024-09-20 04:17:01.452] [Executor task launch worker for task 4.0 in stage 0.0 (TID 4)] [INFO] RssShuffleWriter.writeImpl - Finish write shuffle for appId[local-1726805795493_1726805795466], shuffleId[0], taskId[4_0] with write 944 ms, include checkSendResult[2], commit[0], WriteBufferManager cost copyTime[0], writeTime[938], serializeTime[775], compressTime[64], estimateTime[0], requireMemoryTime[0], uncompressedDataLen[44504032]
[2024-09-20 04:17:01.452] [Grpc-7] [INFO] ShuffleServerGrpcService.reportShuffleResult - Accepted blockIds report for 15 blocks across 5 partitions as shuffle result for task appId[local-1726805795493_1726805795466], shuffleId[0], taskAttemptId[16]
[2024-09-20 04:17:01.452] [Grpc-7] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=reportShuffleResult statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=97 appId=local-1726805795493_1726805795466 shuffleId=0 args{taskAttemptId=16, bitmapNum=1, partitionToBlockIdsSize=5} context{updatedBlockCount=15, expectedBlockCount=15}
[2024-09-20 04:17:01.452] [Executor task launch worker for task 4.0 in stage 0.0 (TID 4)] [INFO] ShuffleWriteClientImpl.reportShuffleResult - Report shuffle result to ShuffleServerInfo{host[10.1.0.33], grpc port[20045]} for appId[local-1726805795493_1726805795466], shuffleId[0] successfully
[2024-09-20 04:17:01.452] [Executor task launch worker for task 4.0 in stage 0.0 (TID 4)] [INFO] RssShuffleWriter.stop - Report shuffle result for task[16] with bitmapNum[1] cost 0 ms
[2024-09-20 04:17:01.453] [Executor task launch worker for task 4.0 in stage 0.0 (TID 4)] [INFO] Executor.logInfo - Finished task 4.0 in stage 0.0 (TID 4). 1644 bytes result sent to driver
[2024-09-20 04:17:01.453] [task-result-getter-0] [INFO] TaskSetManager.logInfo - Finished task 4.0 in stage 0.0 (TID 4) in 1304 ms on fv-az1775-366 (executor driver) (5/5)
[2024-09-20 04:17:01.453] [task-result-getter-0] [INFO] TaskSchedulerImpl.logInfo - Removed TaskSet 0.0, whose tasks have all completed, from pool
[2024-09-20 04:17:01.453] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - ShuffleMapStage 0 (repartition at RepartitionTest.java:97) finished in 25.902 s
[2024-09-20 04:17:01.453] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - looking for newly runnable stages
[2024-09-20 04:17:01.453] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - running: Set()
[2024-09-20 04:17:01.453] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - waiting: Set(ShuffleMapStage 1, ResultStage 2)
[2024-09-20 04:17:01.453] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - failed: Set()
[2024-09-20 04:17:01.454] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - Submitting ShuffleMapStage 1 (MapPartitionsRDD[10] at repartition at RepartitionTest.java:97), which has no missing parents
[2024-09-20 04:17:01.455] [dag-scheduler-event-loop] [INFO] MemoryStore.logInfo - Block broadcast_4 stored as values in memory (estimated size 7.3 KiB, free 2.5 GiB)
[2024-09-20 04:17:01.455] [dag-scheduler-event-loop] [INFO] MemoryStore.logInfo - Block broadcast_4_piece0 stored as bytes in memory (estimated size 4.1 KiB, free 2.5 GiB)
[2024-09-20 04:17:01.456] [dispatcher-BlockManagerMaster] [INFO] BlockManagerInfo.logInfo - Added broadcast_4_piece0 in memory on fv-az1775-366:45159 (size: 4.1 KiB, free: 2.5 GiB)
[2024-09-20 04:17:01.456] [dag-scheduler-event-loop] [INFO] SparkContext.logInfo - Created broadcast 4 from broadcast at DAGScheduler.scala:1535
[2024-09-20 04:17:01.456] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - Submitting 5 missing tasks from ShuffleMapStage 1 (MapPartitionsRDD[10] at repartition at RepartitionTest.java:97) (first 15 tasks are for partitions Vector(0, 1, 2, 3, 4))
[2024-09-20 04:17:01.456] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Adding task set 1.0 with 5 tasks resource profile 0
[2024-09-20 04:17:01.457] [dispatcher-event-loop-3] [INFO] TaskSetManager.logInfo - Starting task 0.0 in stage 1.0 (TID 5) (fv-az1775-366, executor driver, partition 0, ANY, 7446 bytes)
[2024-09-20 04:17:01.457] [dispatcher-event-loop-3] [INFO] TaskSetManager.logInfo - Starting task 1.0 in stage 1.0 (TID 6) (fv-az1775-366, executor driver, partition 1, ANY, 7446 bytes)
[2024-09-20 04:17:01.457] [dispatcher-event-loop-3] [INFO] TaskSetManager.logInfo - Starting task 2.0 in stage 1.0 (TID 7) (fv-az1775-366, executor driver, partition 2, ANY, 7446 bytes)
[2024-09-20 04:17:01.457] [dispatcher-event-loop-3] [INFO] TaskSetManager.logInfo - Starting task 3.0 in stage 1.0 (TID 8) (fv-az1775-366, executor driver, partition 3, ANY, 7446 bytes)
[2024-09-20 04:17:01.457] [Executor task launch worker for task 2.0 in stage 1.0 (TID 7)] [INFO] Executor.logInfo - Running task 2.0 in stage 1.0 (TID 7)
[2024-09-20 04:17:01.457] [Executor task launch worker for task 1.0 in stage 1.0 (TID 6)] [INFO] Executor.logInfo - Running task 1.0 in stage 1.0 (TID 6)
[2024-09-20 04:17:01.457] [Executor task launch worker for task 3.0 in stage 1.0 (TID 8)] [INFO] Executor.logInfo - Running task 3.0 in stage 1.0 (TID 8)
[2024-09-20 04:17:01.458] [Executor task launch worker for task 2.0 in stage 1.0 (TID 7)] [INFO] RssShuffleWriter.<init> - RssShuffle start write taskAttemptId[8] data with RssHandle[appId local-1726805795493_1726805795466, shuffleId 1].
[2024-09-20 04:17:01.458] [Executor task launch worker for task 3.0 in stage 1.0 (TID 8)] [INFO] RssShuffleWriter.<init> - RssShuffle start write taskAttemptId[12] data with RssHandle[appId local-1726805795493_1726805795466, shuffleId 1].
[2024-09-20 04:17:01.460] [Executor task launch worker for task 0.0 in stage 1.0 (TID 5)] [INFO] Executor.logInfo - Running task 0.0 in stage 1.0 (TID 5)
[2024-09-20 04:17:01.460] [Executor task launch worker for task 3.0 in stage 1.0 (TID 8)] [INFO] RssShuffleManager.getReader - Get taskId cost 1 ms, and request expected blockIds from 5 tasks for shuffleId[0], partitionId[3, 4]
[2024-09-20 04:17:01.461] [Executor task launch worker for task 2.0 in stage 1.0 (TID 7)] [INFO] RssShuffleManager.getReader - Get taskId cost 2 ms, and request expected blockIds from 5 tasks for shuffleId[0], partitionId[2, 3]
[2024-09-20 04:17:01.461] [Executor task launch worker for task 1.0 in stage 1.0 (TID 6)] [INFO] RssShuffleWriter.<init> - RssShuffle start write taskAttemptId[4] data with RssHandle[appId local-1726805795493_1726805795466, shuffleId 1].
[2024-09-20 04:17:01.461] [Executor task launch worker for task 0.0 in stage 1.0 (TID 5)] [INFO] RssShuffleWriter.<init> - RssShuffle start write taskAttemptId[0] data with RssHandle[appId local-1726805795493_1726805795466, shuffleId 1].
[2024-09-20 04:17:01.462] [Executor task launch worker for task 0.0 in stage 1.0 (TID 5)] [INFO] RssShuffleManager.getReader - Get taskId cost 0 ms, and request expected blockIds from 5 tasks for shuffleId[0], partitionId[0, 1]
[2024-09-20 04:17:01.462] [Executor task launch worker for task 1.0 in stage 1.0 (TID 6)] [INFO] RssShuffleManager.getReader - Get taskId cost 0 ms, and request expected blockIds from 5 tasks for shuffleId[0], partitionId[1, 2]
[2024-09-20 04:17:01.462] [Grpc-4] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=208 appId=local-1726805795493_1726805795466 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=627} context{bitmap[0].<size,byte>=<455,3366>, partitionBlockCount=91}
[2024-09-20 04:17:01.462] [Grpc-9] [ERROR] ShuffleServerGrpcService.getShuffleResultForMultiPart - Error happened when get shuffle result for appId[local-1726805795493_1726805795466], shuffleId[0], partitions[0]
java.lang.ArrayIndexOutOfBoundsException: 3
at org.roaringbitmap.longlong.Roaring64NavigableMap.ensureOne(Roaring64NavigableMap.java:676)
at org.roaringbitmap.longlong.Roaring64NavigableMap.ensureCumulatives(Roaring64NavigableMap.java:567)
at org.roaringbitmap.longlong.Roaring64NavigableMap.getLongCardinality(Roaring64NavigableMap.java:278)
at org.apache.uniffle.server.ShuffleTaskManager.lambda$getFinishedBlockIds$9(ShuffleTaskManager.java:657)
at java.util.Optional.ifPresent(Optional.java:159)
at org.apache.uniffle.server.ShuffleTaskManager.getFinishedBlockIds(ShuffleTaskManager.java:652)
at org.apache.uniffle.server.ShuffleServerGrpcService.getShuffleResultForMultiPart(ShuffleServerGrpcService.java:985)
at org.apache.uniffle.proto.ShuffleServerGrpc$MethodHandlers.invoke(ShuffleServerGrpc.java:1180)
at io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:182)
at io.grpc.PartialForwardingServerCallListener.onHalfClose(PartialForwardingServerCallListener.java:35)
at io.grpc.ForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:23)
at io.grpc.ForwardingServerCallListener$SimpleForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:40)
at org.apache.uniffle.common.rpc.ClientContextServerInterceptor$1.onHalfClose(ClientContextServerInterceptor.java:63)
at io.grpc.PartialForwardingServerCallListener.onHalfClose(PartialForwardingServerCallListener.java:35)
at io.grpc.ForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:23)
at io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:356)
at io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:861)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:17:01.463] [Grpc-9] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=INTERNAL_ERROR from=/10.1.0.33:42082 executionTimeUs=516 appId=local-1726805795493_1726805795466 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=0}
[2024-09-20 04:17:01.463] [Grpc-6] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=625 appId=local-1726805795493_1726805795466 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=627} context{bitmap[0].<size,byte>=<455,3366>, partitionBlockCount=91}
[2024-09-20 04:17:01.463] [Grpc-0] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=740 appId=local-1726805795493_1726805795466 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=627} context{bitmap[0].<size,byte>=<455,3366>, partitionBlockCount=91}
[2024-09-20 04:17:01.463] [Executor task launch worker for task 3.0 in stage 1.0 (TID 8)] [INFO] RssShuffleManager.getReaderImpl - Get shuffle blockId cost 2 ms, and get 91 blockIds for shuffleId[0], startPartition[3], endPartition[4]
[2024-09-20 04:17:01.463] [Executor task launch worker for task 0.0 in stage 1.0 (TID 5)] [ERROR] ShuffleServerGrpcClient.getShuffleResultForMultiPart - Can't get shuffle result from 10.1.0.33:20045 for [appId=local-1726805795493_1726805795466, shuffleId=0, errorMsg:3
[2024-09-20 04:17:01.463] [Executor task launch worker for task 3.0 in stage 1.0 (TID 8)] [INFO] RssShuffleManager.getReaderImpl - Shuffle reader using remote storage Empty Remote Storage
[2024-09-20 04:17:01.463] [Executor task launch worker for task 1.0 in stage 1.0 (TID 6)] [INFO] RssShuffleManager.getReaderImpl - Get shuffle blockId cost 1 ms, and get 91 blockIds for shuffleId[0], startPartition[1], endPartition[2]
[2024-09-20 04:17:01.464] [Executor task launch worker for task 1.0 in stage 1.0 (TID 6)] [INFO] RssShuffleManager.getReaderImpl - Shuffle reader using remote storage Empty Remote Storage
[2024-09-20 04:17:01.464] [Executor task launch worker for task 0.0 in stage 1.0 (TID 5)] [WARN] ShuffleWriteClientImpl.getShuffleResultForMultiPart - Get shuffle result is failed from ShuffleServerInfo{host[10.1.0.33], grpc port[20045]} for appId[local-1726805795493_1726805795466], shuffleId[0], requestPartitions[0]
org.apache.uniffle.common.exception.RssFetchFailedException: Can't get shuffle result from 10.1.0.33:20045 for [appId=local-1726805795493_1726805795466, shuffleId=0, errorMsg:3
at org.apache.uniffle.client.impl.grpc.ShuffleServerGrpcClient.getShuffleResultForMultiPart(ShuffleServerGrpcClient.java:913)
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:860)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:364)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:328)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:101)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
at org.apache.spark.scheduler.Task.run(Task.scala:139)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:17:01.465] [Executor task launch worker for task 2.0 in stage 1.0 (TID 7)] [INFO] RssShuffleManager.getReaderImpl - Get shuffle blockId cost 4 ms, and get 91 blockIds for shuffleId[0], startPartition[2], endPartition[3]
[2024-09-20 04:17:01.465] [Executor task launch worker for task 0.0 in stage 1.0 (TID 5)] [ERROR] ShuffleWriteClientImpl.getShuffleResultForMultiPart - Failed to meet replica requirement: PartitionDataReplicaRequirementTracking{shuffleId=0, inventory={0={0=[ShuffleServerInfo{host[10.1.0.33], grpc port[20045]}]}, 1={0=[ShuffleServerInfo{host[10.1.0.33], grpc port[20045]}]}, 2={0=[ShuffleServerInfo{host[10.1.0.33], grpc port[20045]}]}, 3={0=[ShuffleServerInfo{host[10.1.0.33], grpc port[20045]}]}, 4={0=[ShuffleServerInfo{host[10.1.0.33], grpc port[20045]}]}}, succeedList={}}
[2024-09-20 04:17:01.465] [Executor task launch worker for task 2.0 in stage 1.0 (TID 7)] [INFO] RssShuffleManager.getReaderImpl - Shuffle reader using remote storage Empty Remote Storage
[2024-09-20 04:17:01.465] [Executor task launch worker for task 0.0 in stage 1.0 (TID 5)] [INFO] RssShuffleManager.markFailedTask - Mark the task: 5_0 failed.
[2024-09-20 04:17:01.465] [Executor task launch worker for task 0.0 in stage 1.0 (TID 5)] [ERROR] Executor.logError - Exception in task 0.0 in stage 1.0 (TID 5)
org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805795493_1726805795466], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:364)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:328)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:101)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
at org.apache.spark.scheduler.Task.run(Task.scala:139)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:17:01.467] [dispatcher-event-loop-0] [INFO] TaskSetManager.logInfo - Starting task 4.0 in stage 1.0 (TID 9) (fv-az1775-366, executor driver, partition 4, ANY, 7446 bytes)
[2024-09-20 04:17:01.467] [Executor task launch worker for task 4.0 in stage 1.0 (TID 9)] [INFO] Executor.logInfo - Running task 4.0 in stage 1.0 (TID 9)
[2024-09-20 04:17:01.467] [task-result-getter-1] [WARN] TaskSetManager.logWarning - Lost task 0.0 in stage 1.0 (TID 5) (fv-az1775-366 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805795493_1726805795466], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:364)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:328)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:101)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
at org.apache.spark.scheduler.Task.run(Task.scala:139)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:17:01.468] [task-result-getter-1] [ERROR] TaskSetManager.logError - Task 0 in stage 1.0 failed 1 times; aborting job
[2024-09-20 04:17:01.468] [Executor task launch worker for task 4.0 in stage 1.0 (TID 9)] [INFO] RssShuffleWriter.<init> - RssShuffle start write taskAttemptId[16] data with RssHandle[appId local-1726805795493_1726805795466, shuffleId 1].
[2024-09-20 04:17:01.469] [Executor task launch worker for task 4.0 in stage 1.0 (TID 9)] [INFO] RssShuffleManager.getReader - Get taskId cost 0 ms, and request expected blockIds from 5 tasks for shuffleId[0], partitionId[4, 5]
[2024-09-20 04:17:01.469] [Grpc-7] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=SUCCESS from=/10.1.0.33:42082 executionTimeUs=99 appId=local-1726805795493_1726805795466 shuffleId=0 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=627} context{bitmap[0].<size,byte>=<455,3366>, partitionBlockCount=91}
[2024-09-20 04:17:01.471] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Cancelling stage 1
[2024-09-20 04:17:01.472] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Killing all running tasks in stage 1: Stage cancelled
[2024-09-20 04:17:01.473] [Executor task launch worker for task 4.0 in stage 1.0 (TID 9)] [INFO] RssShuffleManager.getReaderImpl - Get shuffle blockId cost 3 ms, and get 91 blockIds for shuffleId[0], startPartition[4], endPartition[5]
[2024-09-20 04:17:01.473] [Executor task launch worker for task 4.0 in stage 1.0 (TID 9)] [INFO] RssShuffleManager.getReaderImpl - Shuffle reader using remote storage Empty Remote Storage
[2024-09-20 04:17:01.475] [dispatcher-event-loop-1] [INFO] Executor.logInfo - Executor is trying to kill task 4.0 in stage 1.0 (TID 9), reason: Stage cancelled
[2024-09-20 04:17:01.475] [dispatcher-event-loop-1] [INFO] Executor.logInfo - Executor is trying to kill task 1.0 in stage 1.0 (TID 6), reason: Stage cancelled
[2024-09-20 04:17:01.475] [dispatcher-event-loop-1] [INFO] Executor.logInfo - Executor is trying to kill task 2.0 in stage 1.0 (TID 7), reason: Stage cancelled
[2024-09-20 04:17:01.475] [dispatcher-event-loop-1] [INFO] Executor.logInfo - Executor is trying to kill task 3.0 in stage 1.0 (TID 8), reason: Stage cancelled
[2024-09-20 04:17:01.491] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Stage 1 was cancelled
[2024-09-20 04:17:01.491] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - ShuffleMapStage 1 (repartition at RepartitionTest.java:97) failed in 0.037 s due to Job aborted due to stage failure: Task 0 in stage 1.0 failed 1 times, most recent failure: Lost task 0.0 in stage 1.0 (TID 5) (fv-az1775-366 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Get shuffle result is failed for appId[local-1726805795493_1726805795466], shuffleId[0]
at org.apache.uniffle.client.impl.ShuffleWriteClientImpl.getShuffleResultForMultiPart(ShuffleWriteClientImpl.java:889)
at org.apache.spark.shuffle.RssShuffleManager.getShuffleResultForMultiPart(RssShuffleManager.java:1031)
at org.apache.spark.shuffle.RssShuffleManager.getReaderImpl(RssShuffleManager.java:653)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:586)
at org.apache.spark.shuffle.RssShuffleManager.getReader(RssShuffleManager.java:558)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:364)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:328)
at org.apache.spark.rdd.CoalescedRDD.$anonfun$compute$1(CoalescedRDD.scala:99)
at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
at org.apache.spark.shuffle.writer.RssShuffleWriter.writeImpl(RssShuffleWriter.java:316)
at org.apache.spark.shuffle.writer.RssShuffleWriter.write(RssShuffleWriter.java:287)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:101)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
at org.apache.spark.scheduler.Task.run(Task.scala:139)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Driver stacktrace:
[2024-09-20 04:17:01.493] [main] [INFO] DAGScheduler.logInfo - Job 0 failed: sortByKey at RepartitionTest.java:99, took 25.946739 s
Check failure on line 0 in org.apache.uniffle.test.RepartitionWithHadoopHybridStorageRssTest
github-actions / Test Results
2 out of 10 runs with error: resultCompareTest (org.apache.uniffle.test.RepartitionWithHadoopHybridStorageRssTest)
artifacts/integration-reports-spark3.5-scala2.13/integration-test/spark-common/target/surefire-reports/TEST-org.apache.uniffle.test.RepartitionWithHadoopHybridStorageRssTest.xml [took 20s]
artifacts/integration-reports-spark3/integration-test/spark-common/target/surefire-reports/TEST-org.apache.uniffle.test.RepartitionWithHadoopHybridStorageRssTest.xml [took 27s]
Raw output
Job aborted due to stage failure: Task 0 in stage 5.0 failed 1 times, most recent failure: Lost task 0.0 in stage 5.0 (TID 14) (fv-az1147-284 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Failed to read shuffle data from HOT handler
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:124)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.read(ShuffleReadClientImpl.java:308)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.readShuffleBlockData(ShuffleReadClientImpl.java:216)
at org.apache.spark.shuffle.reader.RssShuffleDataIterator.hasNext(RssShuffleDataIterator.java:116)
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
at org.apache.spark.shuffle.reader.RssShuffleReader$MultiPartitionIterator.hasNext(RssShuffleReader.java:316)
at org.apache.spark.util.collection.ExternalAppendOnlyMap.insertAll(ExternalAppendOnlyMap.scala:155)
at org.apache.spark.Aggregator.combineCombinersByKey(Aggregator.scala:50)
at org.apache.spark.shuffle.reader.RssShuffleReader.read(RssShuffleReader.java:186)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:52)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: io.netty.util.IllegalReferenceCountException: refCnt: 0, decrement: 1
at io.netty.util.internal.ReferenceCountUpdater.toLiveRealRefCnt(ReferenceCountUpdater.java:83)
at io.netty.util.internal.ReferenceCountUpdater.release(ReferenceCountUpdater.java:148)
at io.netty.buffer.AbstractReferenceCountedByteBuf.release(AbstractReferenceCountedByteBuf.java:101)
at org.apache.uniffle.common.netty.buffer.NettyManagedBuffer.release(NettyManagedBuffer.java:59)
at org.apache.uniffle.common.ShuffleIndexResult.release(ShuffleIndexResult.java:77)
at org.apache.uniffle.storage.handler.impl.DataSkippableReadHandler.readShuffleData(DataSkippableReadHandler.java:82)
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:113)
... 21 more
Driver stacktrace:
org.apache.spark.SparkException:
Job aborted due to stage failure: Task 0 in stage 5.0 failed 1 times, most recent failure: Lost task 0.0 in stage 5.0 (TID 14) (fv-az1147-284 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Failed to read shuffle data from HOT handler
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:124)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.read(ShuffleReadClientImpl.java:308)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.readShuffleBlockData(ShuffleReadClientImpl.java:216)
at org.apache.spark.shuffle.reader.RssShuffleDataIterator.hasNext(RssShuffleDataIterator.java:116)
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
at org.apache.spark.shuffle.reader.RssShuffleReader$MultiPartitionIterator.hasNext(RssShuffleReader.java:316)
at org.apache.spark.util.collection.ExternalAppendOnlyMap.insertAll(ExternalAppendOnlyMap.scala:155)
at org.apache.spark.Aggregator.combineCombinersByKey(Aggregator.scala:50)
at org.apache.spark.shuffle.reader.RssShuffleReader.read(RssShuffleReader.java:186)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:52)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: io.netty.util.IllegalReferenceCountException: refCnt: 0, decrement: 1
at io.netty.util.internal.ReferenceCountUpdater.toLiveRealRefCnt(ReferenceCountUpdater.java:83)
at io.netty.util.internal.ReferenceCountUpdater.release(ReferenceCountUpdater.java:148)
at io.netty.buffer.AbstractReferenceCountedByteBuf.release(AbstractReferenceCountedByteBuf.java:101)
at org.apache.uniffle.common.netty.buffer.NettyManagedBuffer.release(NettyManagedBuffer.java:59)
at org.apache.uniffle.common.ShuffleIndexResult.release(ShuffleIndexResult.java:77)
at org.apache.uniffle.storage.handler.impl.DataSkippableReadHandler.readShuffleData(DataSkippableReadHandler.java:82)
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:113)
... 21 more
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2258)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2207)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2206)
at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2206)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1079)
at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1079)
at scala.Option.foreach(Option.scala:407)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1079)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2445)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2387)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2376)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:868)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2196)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2217)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2236)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2261)
at org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1030)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:414)
at org.apache.spark.rdd.RDD.collect(RDD.scala:1029)
at org.apache.spark.rdd.PairRDDFunctions.$anonfun$collectAsMap$1(PairRDDFunctions.scala:737)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:414)
at org.apache.spark.rdd.PairRDDFunctions.collectAsMap(PairRDDFunctions.scala:736)
at org.apache.spark.api.java.JavaPairRDD.collectAsMap(JavaPairRDD.scala:663)
at org.apache.uniffle.test.RepartitionTest.repartitionApp(RepartitionTest.java:99)
at org.apache.uniffle.test.RepartitionTest.runTest(RepartitionTest.java:49)
at org.apache.uniffle.test.SparkIntegrationTestBase.runSparkApp(SparkIntegrationTestBase.java:102)
at org.apache.uniffle.test.SparkIntegrationTestBase.run(SparkIntegrationTestBase.java:68)
at org.apache.uniffle.test.RepartitionTest.resultCompareTest(RepartitionTest.java:44)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725)
at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:107)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:150)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:124)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
Caused by: org.apache.uniffle.common.exception.RssFetchFailedException: Failed to read shuffle data from HOT handler
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:124)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.read(ShuffleReadClientImpl.java:308)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.readShuffleBlockData(ShuffleReadClientImpl.java:216)
at org.apache.spark.shuffle.reader.RssShuffleDataIterator.hasNext(RssShuffleDataIterator.java:116)
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
at org.apache.spark.shuffle.reader.RssShuffleReader$MultiPartitionIterator.hasNext(RssShuffleReader.java:316)
at org.apache.spark.util.collection.ExternalAppendOnlyMap.insertAll(ExternalAppendOnlyMap.scala:155)
at org.apache.spark.Aggregator.combineCombinersByKey(Aggregator.scala:50)
at org.apache.spark.shuffle.reader.RssShuffleReader.read(RssShuffleReader.java:186)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:52)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: io.netty.util.IllegalReferenceCountException: refCnt: 0, decrement: 1
at io.netty.util.internal.ReferenceCountUpdater.toLiveRealRefCnt(ReferenceCountUpdater.java:83)
at io.netty.util.internal.ReferenceCountUpdater.release(ReferenceCountUpdater.java:148)
at io.netty.buffer.AbstractReferenceCountedByteBuf.release(AbstractReferenceCountedByteBuf.java:101)
at org.apache.uniffle.common.netty.buffer.NettyManagedBuffer.release(NettyManagedBuffer.java:59)
at org.apache.uniffle.common.ShuffleIndexResult.release(ShuffleIndexResult.java:77)
at org.apache.uniffle.storage.handler.impl.DataSkippableReadHandler.readShuffleData(DataSkippableReadHandler.java:82)
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:113)
... 21 more
[2024-09-20 04:15:22.541] [main] [INFO] MiniDFSCluster.<init> - starting cluster: numNameNodes=1, numDataNodes=1
Formatting using clusterid: testClusterID
[2024-09-20 04:15:22.542] [main] [INFO] FSEditLog.newInstance - Edit logging is async:true
[2024-09-20 04:15:22.542] [main] [INFO] FSNamesystem.<init> - KeyProvider: null
[2024-09-20 04:15:22.542] [main] [INFO] FSNamesystem.<init> - fsLock is fair: true
[2024-09-20 04:15:22.542] [main] [INFO] FSNamesystem.<init> - Detailed lock hold time metrics enabled: false
[2024-09-20 04:15:22.543] [main] [INFO] DatanodeManager.<init> - dfs.block.invalidate.limit=1000
[2024-09-20 04:15:22.543] [main] [INFO] DatanodeManager.<init> - dfs.namenode.datanode.registration.ip-hostname-check=true
[2024-09-20 04:15:22.543] [main] [INFO] BlockManager.printBlockDeletionTime - dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
[2024-09-20 04:15:22.543] [main] [INFO] BlockManager.printBlockDeletionTime - The block deletion will start around 2024 Sep 20 04:15:22
[2024-09-20 04:15:22.543] [main] [INFO] GSet.computeCapacity - Computing capacity for map BlocksMap
[2024-09-20 04:15:22.543] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:15:22.543] [main] [INFO] GSet.computeCapacity - 2.0% max memory 4.4 GB = 91.0 MB
[2024-09-20 04:15:22.543] [main] [INFO] GSet.computeCapacity - capacity = 2^24 = 16777216 entries
[2024-09-20 04:15:22.546] [main] [INFO] BlockManager.createBlockTokenSecretManager - dfs.block.access.token.enable=false
[2024-09-20 04:15:22.546] [main] [INFO] BlockManager.<init> - defaultReplication = 1
[2024-09-20 04:15:22.546] [main] [INFO] BlockManager.<init> - maxReplication = 512
[2024-09-20 04:15:22.546] [main] [INFO] BlockManager.<init> - minReplication = 1
[2024-09-20 04:15:22.546] [main] [INFO] BlockManager.<init> - maxReplicationStreams = 2
[2024-09-20 04:15:22.546] [main] [INFO] BlockManager.<init> - replicationRecheckInterval = 3000
[2024-09-20 04:15:22.546] [main] [INFO] BlockManager.<init> - encryptDataTransfer = false
[2024-09-20 04:15:22.546] [main] [INFO] BlockManager.<init> - maxNumBlocksToLog = 1000
[2024-09-20 04:15:22.546] [main] [INFO] FSNamesystem.<init> - fsOwner = runner (auth:SIMPLE)
[2024-09-20 04:15:22.546] [main] [INFO] FSNamesystem.<init> - supergroup = supergroup
[2024-09-20 04:15:22.546] [main] [INFO] FSNamesystem.<init> - isPermissionEnabled = true
[2024-09-20 04:15:22.546] [main] [INFO] FSNamesystem.<init> - HA Enabled: false
[2024-09-20 04:15:22.547] [main] [INFO] FSNamesystem.<init> - Append Enabled: true
[2024-09-20 04:15:22.547] [main] [INFO] GSet.computeCapacity - Computing capacity for map INodeMap
[2024-09-20 04:15:22.547] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:15:22.547] [main] [INFO] GSet.computeCapacity - 1.0% max memory 4.4 GB = 45.5 MB
[2024-09-20 04:15:22.547] [main] [INFO] GSet.computeCapacity - capacity = 2^23 = 8388608 entries
[2024-09-20 04:15:22.549] [main] [INFO] FSDirectory.<init> - ACLs enabled? false
[2024-09-20 04:15:22.549] [main] [INFO] FSDirectory.<init> - XAttrs enabled? true
[2024-09-20 04:15:22.549] [main] [INFO] NameNode.<init> - Caching file names occurring more than 10 times
[2024-09-20 04:15:22.549] [main] [INFO] GSet.computeCapacity - Computing capacity for map cachedBlocks
[2024-09-20 04:15:22.549] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:15:22.549] [main] [INFO] GSet.computeCapacity - 0.25% max memory 4.4 GB = 11.4 MB
[2024-09-20 04:15:22.549] [main] [INFO] GSet.computeCapacity - capacity = 2^21 = 2097152 entries
[2024-09-20 04:15:22.550] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.threshold-pct = 0.9990000128746033
[2024-09-20 04:15:22.550] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.min.datanodes = 0
[2024-09-20 04:15:22.550] [main] [INFO] FSNamesystem.<init> - dfs.namenode.safemode.extension = 0
[2024-09-20 04:15:22.550] [main] [INFO] TopMetrics.logConf - NNTop conf: dfs.namenode.top.window.num.buckets = 10
[2024-09-20 04:15:22.550] [main] [INFO] TopMetrics.logConf - NNTop conf: dfs.namenode.top.num.users = 10
[2024-09-20 04:15:22.550] [main] [INFO] TopMetrics.logConf - NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25
[2024-09-20 04:15:22.550] [main] [INFO] FSNamesystem.initRetryCache - Retry cache on namenode is enabled
[2024-09-20 04:15:22.550] [main] [INFO] FSNamesystem.initRetryCache - Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
[2024-09-20 04:15:22.550] [main] [INFO] GSet.computeCapacity - Computing capacity for map NameNodeRetryCache
[2024-09-20 04:15:22.550] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:15:22.551] [main] [INFO] GSet.computeCapacity - 0.029999999329447746% max memory 4.4 GB = 1.4 MB
[2024-09-20 04:15:22.551] [main] [INFO] GSet.computeCapacity - capacity = 2^17 = 131072 entries
[2024-09-20 04:15:22.551] [main] [INFO] FSImage.format - Allocated new BlockPoolId: BP-1107612211-127.0.0.1-1726805722551
[2024-09-20 04:15:22.553] [main] [INFO] Storage.format - Storage directory /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name1 has been successfully formatted.
[2024-09-20 04:15:22.554] [main] [INFO] Storage.format - Storage directory /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name2 has been successfully formatted.
[2024-09-20 04:15:22.554] [FSImageSaver for /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name1 of type IMAGE_AND_EDITS] [INFO] FSImageFormatProtobuf.save - Saving image file /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name1/current/fsimage.ckpt_0000000000000000000 using no compression
[2024-09-20 04:15:22.554] [FSImageSaver for /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name2 of type IMAGE_AND_EDITS] [INFO] FSImageFormatProtobuf.save - Saving image file /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name2/current/fsimage.ckpt_0000000000000000000 using no compression
[2024-09-20 04:15:22.561] [FSImageSaver for /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name2 of type IMAGE_AND_EDITS] [INFO] FSImageFormatProtobuf.save - Image file /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name2/current/fsimage.ckpt_0000000000000000000 of size 323 bytes saved in 0 seconds.
[2024-09-20 04:15:22.561] [FSImageSaver for /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name1 of type IMAGE_AND_EDITS] [INFO] FSImageFormatProtobuf.save - Image file /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/junit840446445042916468/name1/current/fsimage.ckpt_0000000000000000000 of size 323 bytes saved in 0 seconds.
[2024-09-20 04:15:22.562] [main] [INFO] NNStorageRetentionManager.getImageTxIdToRetain - Going to retain 1 images with txid >= 0
[2024-09-20 04:15:22.563] [main] [INFO] NameNode.createNameNode - createNameNode []
[2024-09-20 04:15:22.564] [main] [WARN] MetricsConfig.loadFirst - Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[2024-09-20 04:15:22.565] [main] [INFO] MetricsSystemImpl.startTimer - Scheduled Metric snapshot period at 10 second(s).
[2024-09-20 04:15:22.565] [main] [INFO] MetricsSystemImpl.start - NameNode metrics system started
[2024-09-20 04:15:22.565] [main] [INFO] NameNode.setClientNamenodeAddress - fs.defaultFS is hdfs://127.0.0.1:0
[2024-09-20 04:15:22.569] [org.apache.hadoop.util.JvmPauseMonitor$Monitor@265aa36c] [INFO] JvmPauseMonitor.run - Starting JVM pause monitor
[2024-09-20 04:15:22.569] [main] [INFO] DFSUtil.httpServerTemplateForNNAndJN - Starting Web-server for hdfs at: http://localhost:0
[2024-09-20 04:15:22.570] [main] [INFO] AuthenticationFilter.constructSecretProvider - Unable to initialize FileSignerSecretProvider, falling back to use random secrets.
[2024-09-20 04:15:22.570] [main] [WARN] HttpRequestLog.getRequestLog - Jetty request log can only be enabled using Log4j
[2024-09-20 04:15:22.571] [main] [INFO] HttpServer2.addGlobalFilter - Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
[2024-09-20 04:15:22.571] [main] [INFO] HttpServer2.addFilter - Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context hdfs
[2024-09-20 04:15:22.571] [main] [INFO] HttpServer2.addFilter - Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
[2024-09-20 04:15:22.571] [main] [INFO] HttpServer2.initWebHdfs - Added filter 'org.apache.hadoop.hdfs.web.AuthFilter' (class=org.apache.hadoop.hdfs.web.AuthFilter)
[2024-09-20 04:15:22.572] [main] [INFO] HttpServer2.addJerseyResourcePackage - addJerseyResourcePackage: packageName=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources, pathSpec=/webhdfs/v1/*
[2024-09-20 04:15:22.572] [main] [INFO] HttpServer2.openListeners - Jetty bound to port 44991
[2024-09-20 04:15:22.572] [main] [INFO] log.info - jetty-6.1.26
[2024-09-20 04:15:22.576] [main] [INFO] log.info - Extract jar:file:/home/runner/.m2/repository/org/apache/hadoop/hadoop-hdfs/2.8.5/hadoop-hdfs-2.8.5-tests.jar!/webapps/hdfs to /home/runner/work/incubator-uniffle/incubator-uniffle/integration-test/spark-common/target/tmp/Jetty_localhost_44991_hdfs____alzkfv/webapp
[2024-09-20 04:15:22.639] [main] [INFO] log.info - Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:44991
[2024-09-20 04:15:22.640] [main] [INFO] FSEditLog.newInstance - Edit logging is async:true
[2024-09-20 04:15:22.640] [main] [INFO] FSNamesystem.<init> - KeyProvider: null
[2024-09-20 04:15:22.641] [main] [INFO] FSNamesystem.<init> - fsLock is fair: true
[2024-09-20 04:15:22.641] [main] [INFO] FSNamesystem.<init> - Detailed lock hold time metrics enabled: false
[2024-09-20 04:15:22.641] [main] [INFO] DatanodeManager.<init> - dfs.block.invalidate.limit=1000
[2024-09-20 04:15:22.641] [main] [INFO] DatanodeManager.<init> - dfs.namenode.datanode.registration.ip-hostname-check=true
[2024-09-20 04:15:22.641] [main] [INFO] BlockManager.printBlockDeletionTime - dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
[2024-09-20 04:15:22.641] [main] [INFO] BlockManager.printBlockDeletionTime - The block deletion will start around 2024 Sep 20 04:15:22
[2024-09-20 04:15:22.642] [main] [INFO] GSet.computeCapacity - Computing capacity for map BlocksMap
[2024-09-20 04:15:22.642] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:15:22.642] [main] [INFO] GSet.computeCapacity - 2.0% max memory 4.4 GB = 91.0 MB
[2024-09-20 04:15:22.642] [main] [INFO] GSet.computeCapacity - capacity = 2^24 = 16777216 entries
[2024-09-20 04:15:22.644] [main] [INFO] BlockManager.createBlockTokenSecretManager - dfs.block.access.token.enable=false
[2024-09-20 04:15:22.644] [main] [INFO] BlockManager.<init> - defaultReplication = 1
[2024-09-20 04:15:22.644] [main] [INFO] BlockManager.<init> - maxReplication = 512
[2024-09-20 04:15:22.644] [main] [INFO] BlockManager.<init> - minReplication = 1
[2024-09-20 04:15:22.644] [main] [INFO] BlockManager.<init> - maxReplicationStreams = 2
[2024-09-20 04:15:22.644] [main] [INFO] BlockManager.<init> - replicationRecheckInterval = 3000
[2024-09-20 04:15:22.644] [main] [INFO] BlockManager.<init> - encryptDataTransfer = false
[2024-09-20 04:15:22.644] [main] [INFO] BlockManager.<init> - maxNumBlocksToLog = 1000
[2024-09-20 04:15:22.644] [main] [INFO] FSNamesystem.<init> - fsOwner = runner (auth:SIMPLE)
[2024-09-20 04:15:22.644] [main] [INFO] FSNamesystem.<init> - supergroup = supergroup
[2024-09-20 04:15:22.644] [main] [INFO] FSNamesystem.<init> - isPermissionEnabled = true
[2024-09-20 04:15:22.644] [main] [INFO] FSNamesystem.<init> - HA Enabled: false
[2024-09-20 04:15:22.645] [main] [INFO] FSNamesystem.<init> - Append Enabled: true
[2024-09-20 04:15:22.645] [main] [INFO] GSet.computeCapacity - Computing capacity for map INodeMap
[2024-09-20 04:15:22.645] [main] [INFO] GSet.computeCapacity - VM type = 64-bit
[2024-09-20 04:15:22.645] [main] [INFO] GSet.computeCapacity - 1.0% max memory 4.4 GB = 45.5 MB
[2024-09-20 04:15:22.645] [main] [INFO] GSet.computeCapacity - capacity = 2^23 = 8388608 entries
[2024-09-20 04:15:22.646] [main] [INFO] FSDirectory.<init> - ACLs enabled? false
[2024-09-20 04:15:22.646] [main] [INFO] FSDirectory.<init> - XAttrs enabled? true
[2024-09-20 04:15:22.646] [main] [INFO] NameNode.<init> - Caching file names occurring more than 10 times
[2024-09-20 04:15:22.646] [main] [INFO]…task 3.0 in stage 5.0 (TID 17)] [INFO] RssShuffleReader.read - Shuffle read started:appId=local-1726805745639_1726805745613, shuffleId=1,taskId=17_0, partitions: [3, 4), maps: [0, 2147483647)
[2024-09-20 04:15:52.713] [nioEventLoopGroup-432-1] [WARN] ShuffleServerNettyHandler.handleGetLocalShuffleIndexRequest - Index file for appId[local-1726805745639_1726805745613], shuffleId[1], partitionId[3] is not found, maybe the data has been flushed to cold storage.
org.apache.uniffle.common.exception.FileNotFoundException: No such data in current storage manager.
at org.apache.uniffle.server.ShuffleTaskManager.getShuffleIndex(ShuffleTaskManager.java:757)
at org.apache.uniffle.server.netty.ShuffleServerNettyHandler.handleGetLocalShuffleIndexRequest(ShuffleServerNettyHandler.java:554)
at org.apache.uniffle.server.netty.ShuffleServerNettyHandler.receive(ShuffleServerNettyHandler.java:109)
at org.apache.uniffle.common.netty.handle.TransportRequestHandler.handle(TransportRequestHandler.java:62)
at org.apache.uniffle.common.netty.handle.TransportChannelHandler.channelRead(TransportChannelHandler.java:100)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:289)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.apache.uniffle.common.netty.TransportFrameDecoder.channelRead(TransportFrameDecoder.java:81)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:15:52.714] [nioEventLoopGroup-432-1] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getLocalShuffleIndex statusCode=SUCCESS from=/10.1.0.42:41836 executionTimeUs=430 appId=local-1726805745639_1726805745613 shuffleId=1 args{requestId=1043, partitionId=3, partitionNumPerRange=1, partitionNum=5}
[2024-09-20 04:15:52.714] [Executor task launch worker for task 3.0 in stage 5.0 (TID 17)] [INFO] ShuffleServerGrpcNettyClient.getShuffleIndex - GetShuffleIndex size:0(bytes) from 10.1.0.42:21013 for appId[local-1726805745639_1726805745613], shuffleId[1], partitionId[3] cost:1(ms)
[2024-09-20 04:15:52.725] [Executor task launch worker for task 0.0 in stage 5.0 (TID 14)] [INFO] RssShuffleReader.read - Shuffle read started:appId=local-1726805745639_1726805745613, shuffleId=1,taskId=14_0, partitions: [0, 1), maps: [0, 2147483647)
[2024-09-20 04:15:52.726] [nioEventLoopGroup-432-1] [WARN] ShuffleServerNettyHandler.handleGetLocalShuffleIndexRequest - Index file for appId[local-1726805745639_1726805745613], shuffleId[1], partitionId[0] is not found, maybe the data has been flushed to cold storage.
org.apache.uniffle.common.exception.FileNotFoundException: No such data in current storage manager.
at org.apache.uniffle.server.ShuffleTaskManager.getShuffleIndex(ShuffleTaskManager.java:757)
at org.apache.uniffle.server.netty.ShuffleServerNettyHandler.handleGetLocalShuffleIndexRequest(ShuffleServerNettyHandler.java:554)
at org.apache.uniffle.server.netty.ShuffleServerNettyHandler.receive(ShuffleServerNettyHandler.java:109)
at org.apache.uniffle.common.netty.handle.TransportRequestHandler.handle(TransportRequestHandler.java:62)
at org.apache.uniffle.common.netty.handle.TransportChannelHandler.channelRead(TransportChannelHandler.java:100)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:289)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.apache.uniffle.common.netty.TransportFrameDecoder.channelRead(TransportFrameDecoder.java:81)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:15:52.727] [nioEventLoopGroup-432-1] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getLocalShuffleIndex statusCode=SUCCESS from=/10.1.0.42:41836 executionTimeUs=444 appId=local-1726805745639_1726805745613 shuffleId=1 args{requestId=1044, partitionId=0, partitionNumPerRange=1, partitionNum=5}
[2024-09-20 04:15:52.727] [IPC Server handler 0 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=listStatus src=/rss/test/local-1726805745639_1726805745613/1/1-1 dst=null perm=null proto=rpc
[2024-09-20 04:15:52.729] [Executor task launch worker for task 1.0 in stage 5.0 (TID 15)] [INFO] HadoopClientReadHandler.init - Find index file for shuffleId[1], partitionId[1] hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0.index
[2024-09-20 04:15:52.729] [IPC Server handler 8 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=listStatus src=/rss/test/local-1726805745639_1726805745613/1/3-3 dst=null perm=null proto=rpc
[2024-09-20 04:15:52.729] [Executor task launch worker for task 3.0 in stage 5.0 (TID 17)] [INFO] HadoopClientReadHandler.init - Find index file for shuffleId[1], partitionId[3] hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0.index
[2024-09-20 04:15:52.729] [IPC Server handler 1 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.730] [IPC Server handler 4 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.730] [IPC Server handler 3 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=open src=/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.730] [IPC Server handler 2 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=open src=/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.731] [IPC Server handler 9 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.731] [IPC Server handler 5 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.732] [IPC Server handler 7 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=open src=/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.732] [Executor task launch worker for task 1.0 in stage 5.0 (TID 15)] [INFO] HadoopClientReadHandler.init - Reading order of Hadoop files with name prefix: [hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0]
[2024-09-20 04:15:52.732] [IPC Server handler 6 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=open src=/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.733] [IPC Server handler 0 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.733] [Executor task launch worker for task 3.0 in stage 5.0 (TID 17)] [INFO] HadoopClientReadHandler.init - Reading order of Hadoop files with name prefix: [hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0]
[2024-09-20 04:15:52.733] [IPC Server handler 8 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.734] [IPC Server handler 1 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.734] [Executor task launch worker for task 1.0 in stage 5.0 (TID 15)] [INFO] HadoopShuffleReadHandler.readShuffleIndex - Read index files hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/1-1/10.1.0.42-20039-21013_0_0.index for 2 ms
[2024-09-20 04:15:52.739] [IPC Server handler 4 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.740] [Executor task launch worker for task 2.0 in stage 5.0 (TID 16)] [INFO] RssShuffleReader.read - Shuffle read started:appId=local-1726805745639_1726805745613, shuffleId=1,taskId=16_0, partitions: [2, 3), maps: [0, 2147483647)
[2024-09-20 04:15:52.740] [Executor task launch worker for task 0.0 in stage 5.0 (TID 14)] [INFO] ShuffleServerGrpcNettyClient.getShuffleIndex - GetShuffleIndex size:70(bytes) from 10.1.0.42:21013 for appId[local-1726805745639_1726805745613], shuffleId[1], partitionId[0] cost:14(ms)
[2024-09-20 04:15:52.740] [nioEventLoopGroup-432-1] [WARN] ShuffleServerNettyHandler.handleGetLocalShuffleIndexRequest - Index file for appId[local-1726805745639_1726805745613], shuffleId[1], partitionId[2] is not found, maybe the data has been flushed to cold storage.
org.apache.uniffle.common.exception.FileNotFoundException: No such data in current storage manager.
at org.apache.uniffle.server.ShuffleTaskManager.getShuffleIndex(ShuffleTaskManager.java:757)
at org.apache.uniffle.server.netty.ShuffleServerNettyHandler.handleGetLocalShuffleIndexRequest(ShuffleServerNettyHandler.java:554)
at org.apache.uniffle.server.netty.ShuffleServerNettyHandler.receive(ShuffleServerNettyHandler.java:109)
at org.apache.uniffle.common.netty.handle.TransportRequestHandler.handle(TransportRequestHandler.java:62)
at org.apache.uniffle.common.netty.handle.TransportChannelHandler.channelRead(TransportChannelHandler.java:100)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:289)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at org.apache.uniffle.common.netty.TransportFrameDecoder.channelRead(TransportFrameDecoder.java:81)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:750)
[2024-09-20 04:15:52.740] [nioEventLoopGroup-432-1] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getLocalShuffleIndex statusCode=SUCCESS from=/10.1.0.42:41836 executionTimeUs=434 appId=local-1726805745639_1726805745613 shuffleId=1 args{requestId=1045, partitionId=2, partitionNumPerRange=1, partitionNum=5}
[2024-09-20 04:15:52.741] [Executor task launch worker for task 2.0 in stage 5.0 (TID 16)] [INFO] ShuffleServerGrpcNettyClient.getShuffleIndex - GetShuffleIndex size:0(bytes) from 10.1.0.42:21013 for appId[local-1726805745639_1726805745613], shuffleId[1], partitionId[2] cost:1(ms)
[2024-09-20 04:15:52.741] [Executor task launch worker for task 0.0 in stage 5.0 (TID 14)] [ERROR] Executor.logError - Exception in task 0.0 in stage 5.0 (TID 14)
org.apache.uniffle.common.exception.RssFetchFailedException: Failed to read shuffle data from HOT handler
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:124)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.read(ShuffleReadClientImpl.java:308)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.readShuffleBlockData(ShuffleReadClientImpl.java:216)
at org.apache.spark.shuffle.reader.RssShuffleDataIterator.hasNext(RssShuffleDataIterator.java:116)
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
at org.apache.spark.shuffle.reader.RssShuffleReader$MultiPartitionIterator.hasNext(RssShuffleReader.java:316)
at org.apache.spark.util.collection.ExternalAppendOnlyMap.insertAll(ExternalAppendOnlyMap.scala:155)
at org.apache.spark.Aggregator.combineCombinersByKey(Aggregator.scala:50)
at org.apache.spark.shuffle.reader.RssShuffleReader.read(RssShuffleReader.java:186)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:52)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: io.netty.util.IllegalReferenceCountException: refCnt: 0, decrement: 1
at io.netty.util.internal.ReferenceCountUpdater.toLiveRealRefCnt(ReferenceCountUpdater.java:83)
at io.netty.util.internal.ReferenceCountUpdater.release(ReferenceCountUpdater.java:148)
at io.netty.buffer.AbstractReferenceCountedByteBuf.release(AbstractReferenceCountedByteBuf.java:101)
at org.apache.uniffle.common.netty.buffer.NettyManagedBuffer.release(NettyManagedBuffer.java:59)
at org.apache.uniffle.common.ShuffleIndexResult.release(ShuffleIndexResult.java:77)
at org.apache.uniffle.storage.handler.impl.DataSkippableReadHandler.readShuffleData(DataSkippableReadHandler.java:82)
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:113)
... 21 more
[2024-09-20 04:15:52.750] [dispatcher-event-loop-0] [INFO] TaskSetManager.logInfo - Starting task 4.0 in stage 5.0 (TID 18) (fv-az1147-284, executor driver, partition 4, ANY, 4260 bytes) taskResourceAssignments Map()
[2024-09-20 04:15:52.750] [Executor task launch worker for task 4.0 in stage 5.0 (TID 18)] [INFO] Executor.logInfo - Running task 4.0 in stage 5.0 (TID 18)
[2024-09-20 04:15:52.750] [task-result-getter-2] [WARN] TaskSetManager.logWarning - Lost task 0.0 in stage 5.0 (TID 14) (fv-az1147-284 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Failed to read shuffle data from HOT handler
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:124)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.read(ShuffleReadClientImpl.java:308)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.readShuffleBlockData(ShuffleReadClientImpl.java:216)
at org.apache.spark.shuffle.reader.RssShuffleDataIterator.hasNext(RssShuffleDataIterator.java:116)
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
at org.apache.spark.shuffle.reader.RssShuffleReader$MultiPartitionIterator.hasNext(RssShuffleReader.java:316)
at org.apache.spark.util.collection.ExternalAppendOnlyMap.insertAll(ExternalAppendOnlyMap.scala:155)
at org.apache.spark.Aggregator.combineCombinersByKey(Aggregator.scala:50)
at org.apache.spark.shuffle.reader.RssShuffleReader.read(RssShuffleReader.java:186)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:52)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: io.netty.util.IllegalReferenceCountException: refCnt: 0, decrement: 1
at io.netty.util.internal.ReferenceCountUpdater.toLiveRealRefCnt(ReferenceCountUpdater.java:83)
at io.netty.util.internal.ReferenceCountUpdater.release(ReferenceCountUpdater.java:148)
at io.netty.buffer.AbstractReferenceCountedByteBuf.release(AbstractReferenceCountedByteBuf.java:101)
at org.apache.uniffle.common.netty.buffer.NettyManagedBuffer.release(NettyManagedBuffer.java:59)
at org.apache.uniffle.common.ShuffleIndexResult.release(ShuffleIndexResult.java:77)
at org.apache.uniffle.storage.handler.impl.DataSkippableReadHandler.readShuffleData(DataSkippableReadHandler.java:82)
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:113)
... 21 more
[2024-09-20 04:15:52.751] [task-result-getter-2] [ERROR] TaskSetManager.logError - Task 0 in stage 5.0 failed 1 times; aborting job
[2024-09-20 04:15:52.751] [Executor task launch worker for task 4.0 in stage 5.0 (TID 18)] [INFO] RssShuffleWriter.<init> - RssShuffle start write taskAttemptId[16] data with RssHandle[appId local-1726805745639_1726805745613, shuffleId 2].
[2024-09-20 04:15:52.752] [Executor task launch worker for task 3.0 in stage 5.0 (TID 17)] [INFO] HadoopShuffleReadHandler.readShuffleIndex - Read index files hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/3-3/10.1.0.42-20039-21013_0_0.index for 19 ms
[2024-09-20 04:15:52.753] [Executor task launch worker for task 4.0 in stage 5.0 (TID 18)] [INFO] RssShuffleManager.getReader - Get taskId cost 0 ms, and request expected blockIds from 5 tasks for shuffleId[1], partitionId[4, 5]
[2024-09-20 04:15:52.753] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Cancelling stage 5
[2024-09-20 04:15:52.753] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Killing all running tasks in stage 5: Stage cancelled
[2024-09-20 04:15:52.753] [Grpc-0] [INFO] SHUFFLE_SERVER_RPC_AUDIT_LOG.close - cmd=getShuffleResultForMultiPart statusCode=SUCCESS from=/10.1.0.42:45444 executionTimeUs=81 appId=local-1726805745639_1726805745613 shuffleId=1 args{partitionsListSize=1, blockIdLayout=blockIdLayout[seq: 21 bits, part: 20 bits, task: 22 bits]} return{serializedBlockIdsBytes=35} context{bitmap[0].<size,byte>=<25,172>, partitionBlockCount=5}
[2024-09-20 04:15:52.753] [Executor task launch worker for task 4.0 in stage 5.0 (TID 18)] [INFO] RssShuffleManager.getReaderImpl - Get shuffle blockId cost 0 ms, and get 5 blockIds for shuffleId[1], startPartition[4], endPartition[5]
[2024-09-20 04:15:52.753] [Executor task launch worker for task 4.0 in stage 5.0 (TID 18)] [INFO] RssShuffleManager.getReaderImpl - Shuffle reader using remote storage hdfs://localhost:43853/rss/test,empty conf
[2024-09-20 04:15:52.755] [dispatcher-event-loop-3] [INFO] Executor.logInfo - Executor is trying to kill task 1.0 in stage 5.0 (TID 15), reason: Stage cancelled
[2024-09-20 04:15:52.756] [dispatcher-event-loop-3] [INFO] Executor.logInfo - Executor is trying to kill task 2.0 in stage 5.0 (TID 16), reason: Stage cancelled
[2024-09-20 04:15:52.756] [dispatcher-event-loop-3] [INFO] Executor.logInfo - Executor is trying to kill task 3.0 in stage 5.0 (TID 17), reason: Stage cancelled
[2024-09-20 04:15:52.756] [dispatcher-event-loop-3] [INFO] Executor.logInfo - Executor is trying to kill task 4.0 in stage 5.0 (TID 18), reason: Stage cancelled
[2024-09-20 04:15:52.756] [dag-scheduler-event-loop] [INFO] TaskSchedulerImpl.logInfo - Stage 5 was cancelled
[2024-09-20 04:15:52.756] [dag-scheduler-event-loop] [INFO] DAGScheduler.logInfo - ShuffleMapStage 5 (reduceByKey at RepartitionTest.java:98) failed in 0.105 s due to Job aborted due to stage failure: Task 0 in stage 5.0 failed 1 times, most recent failure: Lost task 0.0 in stage 5.0 (TID 14) (fv-az1147-284 executor driver): org.apache.uniffle.common.exception.RssFetchFailedException: Failed to read shuffle data from HOT handler
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:124)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.read(ShuffleReadClientImpl.java:308)
at org.apache.uniffle.client.impl.ShuffleReadClientImpl.readShuffleBlockData(ShuffleReadClientImpl.java:216)
at org.apache.spark.shuffle.reader.RssShuffleDataIterator.hasNext(RssShuffleDataIterator.java:116)
at org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
at org.apache.spark.shuffle.reader.RssShuffleReader$MultiPartitionIterator.hasNext(RssShuffleReader.java:316)
at org.apache.spark.util.collection.ExternalAppendOnlyMap.insertAll(ExternalAppendOnlyMap.scala:155)
at org.apache.spark.Aggregator.combineCombinersByKey(Aggregator.scala:50)
at org.apache.spark.shuffle.reader.RssShuffleReader.read(RssShuffleReader.java:186)
at org.apache.spark.rdd.ShuffledRDD.compute(ShuffledRDD.scala:106)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)
at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:52)
at org.apache.spark.scheduler.Task.run(Task.scala:131)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: io.netty.util.IllegalReferenceCountException: refCnt: 0, decrement: 1
at io.netty.util.internal.ReferenceCountUpdater.toLiveRealRefCnt(ReferenceCountUpdater.java:83)
at io.netty.util.internal.ReferenceCountUpdater.release(ReferenceCountUpdater.java:148)
at io.netty.buffer.AbstractReferenceCountedByteBuf.release(AbstractReferenceCountedByteBuf.java:101)
at org.apache.uniffle.common.netty.buffer.NettyManagedBuffer.release(NettyManagedBuffer.java:59)
at org.apache.uniffle.common.ShuffleIndexResult.release(ShuffleIndexResult.java:77)
at org.apache.uniffle.storage.handler.impl.DataSkippableReadHandler.readShuffleData(DataSkippableReadHandler.java:82)
at org.apache.uniffle.storage.handler.impl.ComposedClientReadHandler.readShuffleData(ComposedClientReadHandler.java:113)
... 21 more
Driver stacktrace:
[2024-09-20 04:15:52.759] [main] [INFO] DAGScheduler.logInfo - Job 1 failed: collectAsMap at RepartitionTest.java:99, took 0.112711 s
[2024-09-20 04:15:52.762] [IPC Server handler 3 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=listStatus src=/rss/test/local-1726805745639_1726805745613/1/2-2 dst=null perm=null proto=rpc
[2024-09-20 04:15:52.762] [Executor task launch worker for task 2.0 in stage 5.0 (TID 16)] [INFO] HadoopClientReadHandler.init - Find index file for shuffleId[1], partitionId[2] hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0.index
[2024-09-20 04:15:52.768] [IPC Server handler 2 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.769] [IPC Server handler 9 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=open src=/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.770] [IPC Server handler 5 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.770] [IPC Server handler 7 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=open src=/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.771] [Executor task launch worker for task 2.0 in stage 5.0 (TID 16)] [INFO] HadoopClientReadHandler.init - Reading order of Hadoop files with name prefix: [hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0]
[2024-09-20 04:15:52.771] [IPC Server handler 6 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0.index dst=null perm=null proto=rpc
[2024-09-20 04:15:52.773] [IPC Server handler 0 on 43853] [INFO] audit.logAuditMessage - allowed=true ugi=runner (auth:SIMPLE) ip=/127.0.0.1 cmd=getfileinfo src=/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0.data dst=null perm=null proto=rpc
[2024-09-20 04:15:52.773] [Executor task launch worker for task 2.0 in stage 5.0 (TID 16)] [INFO] HadoopShuffleReadHandler.readShuffleIndex - Read index files hdfs://localhost:43853/rss/test/local-1726805745639_1726805745613/1/2-2/10.1.0.42-20039-21013_0_0.index for 2 ms
Check warning on line 0 in org.apache.uniffle.server.buffer.ShuffleBufferManagerTest
github-actions / Test Results
1 out of 15 runs failed: bufferSizeTest (org.apache.uniffle.server.buffer.ShuffleBufferManagerTest)
artifacts/unit-reports-spark3.2.0/server/target/surefire-reports/TEST-org.apache.uniffle.server.buffer.ShuffleBufferManagerTest.xml [took 0s]
Raw output
Computed -5 differs from dummy binary-search index: -4
java.lang.AssertionError: Computed -5 differs from dummy binary-search index: -4
at org.roaringbitmap.longlong.Roaring64NavigableMap.ensureOne(Roaring64NavigableMap.java:642)
at org.roaringbitmap.longlong.Roaring64NavigableMap.ensureCumulatives(Roaring64NavigableMap.java:567)
at org.roaringbitmap.longlong.Roaring64NavigableMap.getLongCardinality(Roaring64NavigableMap.java:278)
at org.apache.uniffle.server.buffer.ShuffleBufferManagerTest.waitForFlush(ShuffleBufferManagerTest.java:659)
at org.apache.uniffle.server.buffer.ShuffleBufferManagerTest.bufferSizeTest(ShuffleBufferManagerTest.java:453)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725)
at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:107)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:150)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:124)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
[2024-09-20 04:09:10.718] [main] [INFO] ShuffleBufferManager.<init> - Init shuffle buffer manager with capacity: 500, read buffer capacity: 954518732.
[2024-09-20 04:09:10.718] [main] [WARN] ReconfigurableConfManager.register - ReconfigurableConfManager is not initialized. The conf of [rss.server.huge-partition.size.threshold] will not be updated.
[2024-09-20 04:09:10.718] [main] [WARN] ReconfigurableConfManager.register - ReconfigurableConfManager is not initialized. The conf of [rss.server.huge-partition.size.hard.limit] will not be updated.
[2024-09-20 04:09:10.720] [LocalStorage-check-0] [INFO] DefaultStorageMediaProvider.getStorageMediaFor - Default storage type provider returns HDD by default
[2024-09-20 04:09:10.720] [main] [INFO] LocalStorageManager.<init> - Succeed to initialize storage paths: [/home/runner/work/incubator-uniffle/incubator-uniffle/server/target/tmp/junit8670150291410682781/data]
[2024-09-20 04:09:10.721] [main] [WARN] ReconfigurableConfManager.register - ReconfigurableConfManager is not initialized. The conf of [rss.server.health.checker.localStorageExecutionTimeoutMS] will not be updated.
[2024-09-20 04:09:10.721] [main] [INFO] DefaultFlushEventHandler.createFlushEventExecutor - CreateFlushPool, poolSize:10, keepAliveTime:120, queueSize:2147483647
[2024-09-20 04:09:10.721] [main] [INFO] DefaultFlushEventHandler.createFlushEventExecutor - CreateFlushPool, poolSize:5, keepAliveTime:120, queueSize:2147483647
[2024-09-20 04:09:10.721] [main] [INFO] ShuffleBufferManager.<init> - Init shuffle buffer manager with capacity: 500, read buffer capacity: 954518732.
[2024-09-20 04:09:10.721] [main] [WARN] ReconfigurableConfManager.register - ReconfigurableConfManager is not initialized. The conf of [rss.server.huge-partition.size.threshold] will not be updated.
[2024-09-20 04:09:10.721] [main] [WARN] ReconfigurableConfManager.register - ReconfigurableConfManager is not initialized. The conf of [rss.server.huge-partition.size.hard.limit] will not be updated.
[2024-09-20 04:09:10.722] [main] [INFO] ShuffleBufferManager.flushIfNecessary - Start to flush with usedMemory[428], preAllocatedSize[0], inFlushSize[0]
[2024-09-20 04:09:10.722] [main] [INFO] ShuffleBufferManager.pickFlushedShuffle - Pick application_shuffleId[bufferSizeTest/1] with 428 bytes
[2024-09-20 04:09:10.722] [main] [INFO] ShuffleBufferManager.pickFlushedShuffle - Finish flush pick with 428 bytes
[2024-09-20 04:09:10.723] [main] [INFO] ShuffleBufferManager.flush - Already picked enough buffers to flush 428 bytes
[2024-09-20 04:09:10.723] [main] [WARN] ShuffleFlushManager.getCommittedBlockIds - Unexpected value when getCommittedBlockIds for appId[bufferSizeTest]
[2024-09-20 04:09:10.724] [LocalFileFlushEventThreadPool-0] [WARN] ShuffleFlushManager.getMaxConcurrencyPerPartitionWrite - Should not happen that shuffle task info of bufferSizeTest is null.
[2024-09-20 04:09:10.725] [LocalFileFlushEventThreadPool-0] [INFO] LocalStorageMeta.createMetadataIfNotExist - Create metadata of shuffle bufferSizeTest/1.
[2024-09-20 04:09:11.324] [main] [INFO] ShuffleBufferManager.flushIfNecessary - Start to flush with usedMemory[480], preAllocatedSize[0], inFlushSize[0]
[2024-09-20 04:09:11.325] [main] [INFO] ShuffleBufferManager.pickFlushedShuffle - Pick application_shuffleId[bufferSizeTest/1] with 480 bytes
[2024-09-20 04:09:11.325] [main] [INFO] ShuffleBufferManager.pickFlushedShuffle - Finish flush pick with 480 bytes
[2024-09-20 04:09:11.326] [LocalFileFlushEventThreadPool-1] [WARN] ShuffleFlushManager.getMaxConcurrencyPerPartitionWrite - Should not happen that shuffle task info of bufferSizeTest is null.
[2024-09-20 04:09:11.326] [LocalFileFlushEventThreadPool-3] [WARN] ShuffleFlushManager.getMaxConcurrencyPerPartitionWrite - Should not happen that shuffle task info of bufferSizeTest is null.
[2024-09-20 04:09:11.326] [LocalFileFlushEventThreadPool-2] [WARN] ShuffleFlushManager.getMaxConcurrencyPerPartitionWrite - Should not happen that shuffle task info of bufferSizeTest is null.
[2024-09-20 04:09:11.327] [main] [INFO] ShuffleBufferManager.flush - Already picked enough buffers to flush 384 bytes
[2024-09-20 04:09:11.329] [LocalFileFlushEventThreadPool-4] [WARN] ShuffleFlushManager.getMaxConcurrencyPerPartitionWrite - Should not happen that shuffle task info of bufferSizeTest is null.
Check notice on line 0 in .github
github-actions / Test Results
1 skipped test found
There is 1 skipped test, see "Raw output" for the name of the skipped test.
Raw output
org.apache.uniffle.test.ShuffleServerGrpcTest ‑ rpcMetricsTest
Check notice on line 0 in .github
github-actions / Test Results
1019 tests found (test 1 to 733)
There are 1019 tests, see "Raw output" for the list of tests 1 to 733.
Raw output
org.apache.hadoop.mapred.SortWriteBufferManagerTest ‑ testCombineBuffer
org.apache.hadoop.mapred.SortWriteBufferManagerTest ‑ testCommitBlocksWhenMemoryShuffleDisabled
org.apache.hadoop.mapred.SortWriteBufferManagerTest ‑ testOnePartition
org.apache.hadoop.mapred.SortWriteBufferManagerTest ‑ testWriteException
org.apache.hadoop.mapred.SortWriteBufferManagerTest ‑ testWriteNormal
org.apache.hadoop.mapred.SortWriteBufferTest ‑ testReadWrite
org.apache.hadoop.mapred.SortWriteBufferTest ‑ testSortBufferIterator
org.apache.hadoop.mapreduce.RssMRUtilsTest ‑ applyDynamicClientConfTest
org.apache.hadoop.mapreduce.RssMRUtilsTest ‑ baskAttemptIdTest
org.apache.hadoop.mapreduce.RssMRUtilsTest ‑ blockConvertTest
org.apache.hadoop.mapreduce.RssMRUtilsTest ‑ partitionIdConvertBlockTest
org.apache.hadoop.mapreduce.RssMRUtilsTest ‑ testEstimateTaskConcurrency
org.apache.hadoop.mapreduce.RssMRUtilsTest ‑ testGetRequiredShuffleServerNumber
org.apache.hadoop.mapreduce.RssMRUtilsTest ‑ testValidateRssClientConf
org.apache.hadoop.mapreduce.task.reduce.EventFetcherTest ‑ extraEventFetch
org.apache.hadoop.mapreduce.task.reduce.EventFetcherTest ‑ missingEventFetch
org.apache.hadoop.mapreduce.task.reduce.EventFetcherTest ‑ multiPassEventFetch
org.apache.hadoop.mapreduce.task.reduce.EventFetcherTest ‑ obsoletedAndTipFailedEventFetch
org.apache.hadoop.mapreduce.task.reduce.EventFetcherTest ‑ singlePassEventFetch
org.apache.hadoop.mapreduce.task.reduce.EventFetcherTest ‑ singlePassWithRepeatedSuccessEventFetch
org.apache.hadoop.mapreduce.task.reduce.FetcherTest ‑ testCodecIsDuplicated
org.apache.hadoop.mapreduce.task.reduce.FetcherTest ‑ writeAndReadDataMergeFailsTestWithRss
org.apache.hadoop.mapreduce.task.reduce.FetcherTest ‑ writeAndReadDataTestWithRss
org.apache.hadoop.mapreduce.task.reduce.FetcherTest ‑ writeAndReadDataTestWithoutRss
org.apache.hadoop.mapreduce.task.reduce.RssInMemoryRemoteMergerTest ‑ mergerTest{File}
org.apache.hadoop.mapreduce.task.reduce.RssRemoteMergeManagerTest ‑ mergerTest{File}
org.apache.spark.shuffle.DelegationRssShuffleManagerTest ‑ testCreateFallback
org.apache.spark.shuffle.DelegationRssShuffleManagerTest ‑ testCreateInDriver
org.apache.spark.shuffle.DelegationRssShuffleManagerTest ‑ testCreateInDriverDenied
org.apache.spark.shuffle.DelegationRssShuffleManagerTest ‑ testCreateInExecutor
org.apache.spark.shuffle.DelegationRssShuffleManagerTest ‑ testTryAccessCluster
org.apache.spark.shuffle.FunctionUtilsTests ‑ testOnceFunction0
org.apache.spark.shuffle.RssShuffleManagerTest ‑ testCreateShuffleManagerServer
org.apache.spark.shuffle.RssShuffleManagerTest ‑ testGetDataDistributionType
org.apache.spark.shuffle.RssShuffleManagerTest ‑ testRssShuffleManagerInterface
org.apache.spark.shuffle.RssShuffleManagerTest ‑ testRssShuffleManagerRegisterShuffle{int}[1]
org.apache.spark.shuffle.RssShuffleManagerTest ‑ testRssShuffleManagerRegisterShuffle{int}[2]
org.apache.spark.shuffle.RssShuffleManagerTest ‑ testRssShuffleManagerRegisterShuffle{int}[3]
org.apache.spark.shuffle.RssShuffleManagerTest ‑ testWithStageRetry
org.apache.spark.shuffle.RssSpark2ShuffleUtilsTest ‑ testCreateFetchFailedException
org.apache.spark.shuffle.RssSpark2ShuffleUtilsTest ‑ testIsStageResubmitSupported
org.apache.spark.shuffle.RssSpark3ShuffleUtilsTest ‑ testCreateFetchFailedException
org.apache.spark.shuffle.RssSpark3ShuffleUtilsTest ‑ testIsStageResubmitSupported
org.apache.spark.shuffle.RssSparkShuffleUtilsTest ‑ applyDynamicClientConfTest
org.apache.spark.shuffle.RssSparkShuffleUtilsTest ‑ odfsConfigurationTest
org.apache.spark.shuffle.RssSparkShuffleUtilsTest ‑ testAssignmentTags
org.apache.spark.shuffle.RssSparkShuffleUtilsTest ‑ testEstimateTaskConcurrency
org.apache.spark.shuffle.RssSparkShuffleUtilsTest ‑ testGetRequiredShuffleServerNumber
org.apache.spark.shuffle.RssSparkShuffleUtilsTest ‑ testValidateRssClientConf
org.apache.spark.shuffle.SparkVersionUtilsTest ‑ testSpark3Version
org.apache.spark.shuffle.SparkVersionUtilsTest ‑ testSparkVersion
org.apache.spark.shuffle.handle.MutableShuffleHandleInfoTest ‑ testCreatePartitionReplicaTracking
org.apache.spark.shuffle.handle.MutableShuffleHandleInfoTest ‑ testListAllPartitionAssignmentServers
org.apache.spark.shuffle.handle.MutableShuffleHandleInfoTest ‑ testUpdateAssignment
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ cleanup
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ readTest1{BlockIdLayout}[1]
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ readTest1{BlockIdLayout}[2]
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ readTest2
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ readTest3
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ readTest4
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ readTest5
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ readTest7
org.apache.spark.shuffle.reader.RssShuffleDataIteratorTest ‑ readTestUncompressedShuffle
org.apache.spark.shuffle.reader.RssShuffleReaderTest ‑ readTest
org.apache.spark.shuffle.writer.DataPusherTest ‑ testSendData
org.apache.spark.shuffle.writer.RssShuffleWriterTest ‑ blockFailureResendTest
org.apache.spark.shuffle.writer.RssShuffleWriterTest ‑ checkBlockSendResultTest
org.apache.spark.shuffle.writer.RssShuffleWriterTest ‑ dataConsistencyWhenSpillTriggeredTest
org.apache.spark.shuffle.writer.RssShuffleWriterTest ‑ postBlockEventTest
org.apache.spark.shuffle.writer.RssShuffleWriterTest ‑ reassignMultiTimesForOnePartitionIdTest
org.apache.spark.shuffle.writer.RssShuffleWriterTest ‑ refreshAssignmentTest
org.apache.spark.shuffle.writer.RssShuffleWriterTest ‑ writeTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ addFirstRecordWithLargeSizeTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ addHugeRecordTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ addNullValueRecordTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ addPartitionDataTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ addRecordCompressedTest{BlockIdLayout}[1]
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ addRecordCompressedTest{BlockIdLayout}[2]
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ addRecordUnCompressedTest{BlockIdLayout}[1]
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ addRecordUnCompressedTest{BlockIdLayout}[2]
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ buildBlockEventsTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ createBlockIdTest{BlockIdLayout}[1]
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ createBlockIdTest{BlockIdLayout}[2]
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ spillByOthersTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ spillByOwnTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ spillByOwnWithSparkTaskMemoryManagerTest
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ spillPartial
org.apache.spark.shuffle.writer.WriteBufferManagerTest ‑ testClearWithSpillRatio
org.apache.spark.shuffle.writer.WriteBufferTest ‑ test
org.apache.tez.common.GetShuffleServerRequestTest ‑ testSerDe
org.apache.tez.common.GetShuffleServerResponseTest ‑ testSerDe
org.apache.tez.common.IdUtilsTest ‑ testConvertTezTaskAttemptID
org.apache.tez.common.InputContextUtilsTest ‑ testGetTezTaskAttemptID
org.apache.tez.common.RssTezUtilsTest ‑ attemptTaskIdTest
org.apache.tez.common.RssTezUtilsTest ‑ baskAttemptIdTest
org.apache.tez.common.RssTezUtilsTest ‑ blockConvertTest
org.apache.tez.common.RssTezUtilsTest ‑ testApplyDynamicClientConf
org.apache.tez.common.RssTezUtilsTest ‑ testComputeShuffleId
org.apache.tez.common.RssTezUtilsTest ‑ testEstimateTaskConcurrency
org.apache.tez.common.RssTezUtilsTest ‑ testFilterRssConf
org.apache.tez.common.RssTezUtilsTest ‑ testGetRequiredShuffleServerNumber
org.apache.tez.common.RssTezUtilsTest ‑ testParseDagId
org.apache.tez.common.RssTezUtilsTest ‑ testParseRssWorker
org.apache.tez.common.RssTezUtilsTest ‑ testPartitionIdConvertBlock
org.apache.tez.common.RssTezUtilsTest ‑ testTaskIdStrToTaskId
org.apache.tez.common.ShuffleAssignmentsInfoWritableTest ‑ testSerDe
org.apache.tez.common.TezIdHelperTest ‑ testTetTaskAttemptId
org.apache.tez.dag.app.RssDAGAppMasterTest ‑ testDagStateChangeCallback
org.apache.tez.dag.app.RssDAGAppMasterTest ‑ testFetchRemoteStorageFromCoordinator{String}[1]
org.apache.tez.dag.app.RssDAGAppMasterTest ‑ testFetchRemoteStorageFromCoordinator{String}[2]
org.apache.tez.dag.app.RssDAGAppMasterTest ‑ testFetchRemoteStorageFromDynamicConf{String}[1]
org.apache.tez.dag.app.RssDAGAppMasterTest ‑ testFetchRemoteStorageFromDynamicConf{String}[2]
org.apache.tez.dag.app.TezRemoteShuffleManagerTest ‑ testTezRemoteShuffleManager
org.apache.tez.dag.app.TezRemoteShuffleManagerTest ‑ testTezRemoteShuffleManagerSecure
org.apache.tez.runtime.library.common.shuffle.impl.RssShuffleManagerTest ‑ testFetchFailed
org.apache.tez.runtime.library.common.shuffle.impl.RssShuffleManagerTest ‑ testProgressWithEmptyPendingHosts
org.apache.tez.runtime.library.common.shuffle.impl.RssShuffleManagerTest ‑ testUseSharedExecutor
org.apache.tez.runtime.library.common.shuffle.impl.RssSimpleFetchedInputAllocatorTest ‑ testAllocate{File}
org.apache.tez.runtime.library.common.shuffle.impl.RssTezFetcherTest ‑ testReadWithDiskFetchedInput{File}
org.apache.tez.runtime.library.common.shuffle.impl.RssTezFetcherTest ‑ testReadWithRemoteFetchedInput{File}
org.apache.tez.runtime.library.common.shuffle.impl.RssTezFetcherTest ‑ writeAndReadDataTestWithoutRss
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssInMemoryMergerTest ‑ mergerTest
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssMergeManagerTest ‑ mergerTest
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testPenalty
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testProgressDuringGetHostWait
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testReducerHealth1
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testReducerHealth2
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testReducerHealth3
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testReducerHealth4
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testReducerHealth5
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testReducerHealth6
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testReducerHealth7
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleSchedulerTest ‑ testShutdownWithInterrupt
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleTest ‑ testKillSelf
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssShuffleTest ‑ testSchedulerTerminatesOnException
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssTezBypassWriterTest ‑ testCalcChecksum
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssTezBypassWriterTest ‑ testWrite
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssTezBypassWriterTest ‑ testWriteDiskFetchInput{File}
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssTezBypassWriterTest ‑ testWriteRemoteFetchInput
org.apache.tez.runtime.library.common.shuffle.orderedgrouped.RssTezShuffleDataFetcherTest ‑ testIteratorWithInMemoryReader
org.apache.tez.runtime.library.common.sort.buffer.WriteBufferManagerTest ‑ testCommitBlocksWhenMemoryShuffleDisabled{File}
org.apache.tez.runtime.library.common.sort.buffer.WriteBufferManagerTest ‑ testFailFastWhenFailedToSendBlocks{File}
org.apache.tez.runtime.library.common.sort.buffer.WriteBufferManagerTest ‑ testWriteException{File}
org.apache.tez.runtime.library.common.sort.buffer.WriteBufferManagerTest ‑ testWriteNormal{File}
org.apache.tez.runtime.library.common.sort.buffer.WriteBufferTest ‑ testReadWrite
org.apache.tez.runtime.library.common.sort.impl.RssSorterTest ‑ testCollectAndRecordsPerPartition
org.apache.tez.runtime.library.common.sort.impl.RssTezPerPartitionRecordTest ‑ testNumPartitions
org.apache.tez.runtime.library.common.sort.impl.RssTezPerPartitionRecordTest ‑ testRssTezIndexHasData
org.apache.tez.runtime.library.common.sort.impl.RssUnSorterTest ‑ testCollectAndRecordsPerPartition
org.apache.tez.runtime.library.input.RssOrderedGroupedKVInputTest ‑ testInterruptWhileAwaitingInput
org.apache.tez.runtime.library.input.RssSortedGroupedMergedInputTest ‑ testSimpleConcatenatedMergedKeyValueInput
org.apache.tez.runtime.library.input.RssSortedGroupedMergedInputTest ‑ testSimpleConcatenatedMergedKeyValuesInput
org.apache.tez.runtime.library.output.RssOrderedPartitionedKVOutputTest ‑ testClose
org.apache.tez.runtime.library.output.RssOrderedPartitionedKVOutputTest ‑ testNonStartedOutput
org.apache.tez.runtime.library.output.RssUnorderedKVOutputTest ‑ testClose
org.apache.tez.runtime.library.output.RssUnorderedKVOutputTest ‑ testNonStartedOutput
org.apache.tez.runtime.library.output.RssUnorderedPartitionedKVOutputTest ‑ testClose
org.apache.tez.runtime.library.output.RssUnorderedPartitionedKVOutputTest ‑ testNonStartedOutput
org.apache.uniffle.cli.AdminRestApiTest ‑ testRunRefreshAccessChecker
org.apache.uniffle.cli.CLIContentUtilsTest ‑ testTableFormat
org.apache.uniffle.cli.UniffleTestAdminCLI ‑ testAdminRefreshCLI
org.apache.uniffle.cli.UniffleTestAdminCLI ‑ testMissingClientCLI
org.apache.uniffle.cli.UniffleTestCLI ‑ testExampleCLI
org.apache.uniffle.cli.UniffleTestCLI ‑ testHelp
org.apache.uniffle.client.ClientUtilsTest ‑ testGenerateTaskIdBitMap
org.apache.uniffle.client.ClientUtilsTest ‑ testGetMaxAttemptNo
org.apache.uniffle.client.ClientUtilsTest ‑ testGetNumberOfSignificantBits
org.apache.uniffle.client.ClientUtilsTest ‑ testValidateClientType
org.apache.uniffle.client.ClientUtilsTest ‑ testWaitUntilDoneOrFail
org.apache.uniffle.client.PartitionDataReplicaRequirementTrackingTest ‑ testMultipleReplicaWithMultiServers
org.apache.uniffle.client.PartitionDataReplicaRequirementTrackingTest ‑ testMultipleReplicaWithSingleServer
org.apache.uniffle.client.PartitionDataReplicaRequirementTrackingTest ‑ testSingleReplicaWithMultiServers
org.apache.uniffle.client.PartitionDataReplicaRequirementTrackingTest ‑ testSingleReplicaWithSingleShuffleServer
org.apache.uniffle.client.factory.ShuffleManagerClientFactoryTest ‑ createShuffleManagerClient
org.apache.uniffle.client.impl.FailedBlockSendTrackerTest ‑ test
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest1
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest10
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest11
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest12
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest13
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest13b
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest14
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest15
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest16
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest2
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest3
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest4
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest5
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest7
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest8
org.apache.uniffle.client.impl.ShuffleReadClientImplTest ‑ readTest9
org.apache.uniffle.client.impl.ShuffleWriteClientImplTest ‑ testAbandonEventWhenTaskFailed
org.apache.uniffle.client.impl.ShuffleWriteClientImplTest ‑ testGetShuffleResult{BlockIdLayout}[1]
org.apache.uniffle.client.impl.ShuffleWriteClientImplTest ‑ testGetShuffleResult{BlockIdLayout}[2]
org.apache.uniffle.client.impl.ShuffleWriteClientImplTest ‑ testRegisterAndUnRegisterShuffleServer
org.apache.uniffle.client.impl.ShuffleWriteClientImplTest ‑ testSendData
org.apache.uniffle.client.impl.ShuffleWriteClientImplTest ‑ testSendDataWithDefectiveServers
org.apache.uniffle.client.impl.ShuffleWriteClientImplTest ‑ testSettingRssClientConfigs
org.apache.uniffle.client.record.reader.BufferedSegmentTest ‑ testMergeResolvedSegmentWithHook
org.apache.uniffle.client.record.reader.RMRecordsReaderTest ‑ testNormalReadWithCombine{String}[1]
org.apache.uniffle.client.record.reader.RMRecordsReaderTest ‑ testNormalReadWithCombine{String}[2]
org.apache.uniffle.client.record.reader.RMRecordsReaderTest ‑ testNormalReadWithoutCombine{String}[1]
org.apache.uniffle.client.record.reader.RMRecordsReaderTest ‑ testNormalReadWithoutCombine{String}[2]
org.apache.uniffle.client.record.reader.RMRecordsReaderTest ‑ testReadMulitPartitionWithCombine{String}[1]
org.apache.uniffle.client.record.reader.RMRecordsReaderTest ‑ testReadMulitPartitionWithCombine{String}[2]
org.apache.uniffle.client.record.reader.RMRecordsReaderTest ‑ testReadMulitPartitionWithoutCombine{String}[1]
org.apache.uniffle.client.record.reader.RMRecordsReaderTest ‑ testReadMulitPartitionWithoutCombine{String}[2]
org.apache.uniffle.client.record.writer.RecordCollectionTest ‑ testSortAndSerializeRecords{String}[1]
org.apache.uniffle.client.record.writer.RecordCollectionTest ‑ testSortCombineAndSerializeRecords{String}[1]
org.apache.uniffle.common.ArgumentsTest ‑ argEmptyTest
org.apache.uniffle.common.ArgumentsTest ‑ argTest
org.apache.uniffle.common.BufferSegmentTest ‑ testEquals
org.apache.uniffle.common.BufferSegmentTest ‑ testGetOffset
org.apache.uniffle.common.BufferSegmentTest ‑ testNotEquals{long, long, int, int, long, long}[1]
org.apache.uniffle.common.BufferSegmentTest ‑ testNotEquals{long, long, int, int, long, long}[2]
org.apache.uniffle.common.BufferSegmentTest ‑ testNotEquals{long, long, int, int, long, long}[3]
org.apache.uniffle.common.BufferSegmentTest ‑ testNotEquals{long, long, int, int, long, long}[4]
org.apache.uniffle.common.BufferSegmentTest ‑ testNotEquals{long, long, int, int, long, long}[5]
org.apache.uniffle.common.BufferSegmentTest ‑ testNotEquals{long, long, int, int, long, long}[6]
org.apache.uniffle.common.BufferSegmentTest ‑ testToString
org.apache.uniffle.common.PartitionRangeTest ‑ testCompareTo
org.apache.uniffle.common.PartitionRangeTest ‑ testEquals
org.apache.uniffle.common.PartitionRangeTest ‑ testHashCode
org.apache.uniffle.common.PartitionRangeTest ‑ testPartitionRange
org.apache.uniffle.common.PartitionRangeTest ‑ testToString
org.apache.uniffle.common.ReconfigurableConfManagerTest ‑ test
org.apache.uniffle.common.ReconfigurableConfManagerTest ‑ testWithoutInitialization
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testEmptyStoragePath{String}[1]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testEmptyStoragePath{String}[2]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testEquals
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testHashCode
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testNotEquals{String}[1]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testNotEquals{String}[2]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testNotEquals{String}[3]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testRemoteStorageInfo{String, Map, String}[1]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testRemoteStorageInfo{String, Map, String}[2]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testRemoteStorageInfo{String, Map, String}[3]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testUncommonConfString{String}[1]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testUncommonConfString{String}[2]
org.apache.uniffle.common.RemoteStorageInfoTest ‑ testUncommonConfString{String}[3]
org.apache.uniffle.common.ServerStatusTest ‑ test
org.apache.uniffle.common.ShuffleBlockInfoTest ‑ testToString
org.apache.uniffle.common.ShuffleDataResultTest ‑ testEmpty
org.apache.uniffle.common.ShuffleIndexResultTest ‑ testEmpty
org.apache.uniffle.common.ShufflePartitionedBlockTest ‑ shufflePartitionedBlockTest
org.apache.uniffle.common.ShufflePartitionedBlockTest ‑ testEquals
org.apache.uniffle.common.ShufflePartitionedBlockTest ‑ testNotEquals{int, long, long, int}[1]
org.apache.uniffle.common.ShufflePartitionedBlockTest ‑ testNotEquals{int, long, long, int}[2]
org.apache.uniffle.common.ShufflePartitionedBlockTest ‑ testNotEquals{int, long, long, int}[3]
org.apache.uniffle.common.ShufflePartitionedBlockTest ‑ testNotEquals{int, long, long, int}[4]
org.apache.uniffle.common.ShufflePartitionedBlockTest ‑ testSize
org.apache.uniffle.common.ShufflePartitionedBlockTest ‑ testToString
org.apache.uniffle.common.ShufflePartitionedDataTest ‑ testToString
org.apache.uniffle.common.ShuffleRegisterInfoTest ‑ testEquals
org.apache.uniffle.common.ShuffleRegisterInfoTest ‑ testToString
org.apache.uniffle.common.ShuffleServerInfoTest ‑ testEquals
org.apache.uniffle.common.ShuffleServerInfoTest ‑ testToString
org.apache.uniffle.common.UnionKeyTest ‑ test
org.apache.uniffle.common.compression.CompressionTest ‑ checkDecompressBufferOffsets
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[10]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[11]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[12]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[13]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[14]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[15]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[16]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[17]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[18]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[19]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[1]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[20]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[21]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[22]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[23]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[24]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[2]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[3]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[4]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[5]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[6]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[7]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[8]
org.apache.uniffle.common.compression.CompressionTest ‑ testCompression{int, Type}[9]
org.apache.uniffle.common.config.ConfigOptionTest ‑ testBasicTypes
org.apache.uniffle.common.config.ConfigOptionTest ‑ testDeprecatedAndFallbackKeys
org.apache.uniffle.common.config.ConfigOptionTest ‑ testDeprecatedKeys
org.apache.uniffle.common.config.ConfigOptionTest ‑ testEnumType
org.apache.uniffle.common.config.ConfigOptionTest ‑ testFallbackKeys
org.apache.uniffle.common.config.ConfigOptionTest ‑ testListTypes
org.apache.uniffle.common.config.ConfigOptionTest ‑ testSetKVWithStringTypeDirectly
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[10]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[11]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[12]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[13]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[8]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToBoolean{Object, Boolean}[9]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[10]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[11]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[12]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[13]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[8]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToDouble{Object, Double}[9]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[10]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[11]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[12]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[13]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[14]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[15]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[16]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[17]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[8]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToFloat{Object, Float}[9]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[8]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToInt{Object, Integer}[9]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[10]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[11]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[12]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[13]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[8]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToLong{Object, Long}[9]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[10]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[11]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[12]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[13]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[14]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[8]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToSizeInBytes{Object, long}[9]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToString{Object, String}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToString{Object, String}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertToString{Object, String}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertValueWithUnsupportedType
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertValue{Object, Class}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertValue{Object, Class}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertValue{Object, Class}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertValue{Object, Class}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertValue{Object, Class}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertValue{Object, Class}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testConvertValue{Object, Class}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testGetAllConfigOptions
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testNonNegativeLongValidator{long}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testNonNegativeLongValidator{long}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testNonNegativeLongValidator{long}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testNonNegativeLongValidator{long}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testNonNegativeLongValidator{long}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testNonNegativeLongValidator{long}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[8]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPercentageDoubleValidator{double}[9]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator2{int}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator2{int}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator2{int}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator2{int}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator2{int}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator2{int}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator{long}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator{long}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator{long}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator{long}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator{long}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator{long}[6]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator{long}[7]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveIntegerValidator{long}[8]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveLongValidator{long}[1]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveLongValidator{long}[2]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveLongValidator{long}[3]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveLongValidator{long}[4]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveLongValidator{long}[5]
org.apache.uniffle.common.config.ConfigUtilsTest ‑ testPositiveLongValidator{long}[6]
org.apache.uniffle.common.config.RssConfTest ‑ testOptionWithDefault
org.apache.uniffle.common.config.RssConfTest ‑ testOptionWithNoDefault
org.apache.uniffle.common.config.RssConfTest ‑ testSetStringAndGetConcreteType
org.apache.uniffle.common.filesystem.HadoopFilesystemProviderTest ‑ testGetSecuredFilesystem
org.apache.uniffle.common.filesystem.HadoopFilesystemProviderTest ‑ testGetSecuredFilesystemButNotInitializeHadoopSecurityContext
org.apache.uniffle.common.filesystem.HadoopFilesystemProviderTest ‑ testWriteAndReadBySecuredFilesystem
org.apache.uniffle.common.future.CompletableFutureExtensionTest ‑ timeoutExceptionTest
org.apache.uniffle.common.merger.MergerTest ‑ testMergeSegmentToFile{String, File}[1]
org.apache.uniffle.common.metrics.MetricReporterFactoryTest ‑ testGetMetricReporter
org.apache.uniffle.common.metrics.MetricsManagerTest ‑ testMetricsManager
org.apache.uniffle.common.metrics.prometheus.PrometheusPushGatewayMetricReporterTest ‑ test
org.apache.uniffle.common.metrics.prometheus.PrometheusPushGatewayMetricReporterTest ‑ testParseGroupingKey
org.apache.uniffle.common.metrics.prometheus.PrometheusPushGatewayMetricReporterTest ‑ testParseIncompleteGroupingKey
org.apache.uniffle.common.netty.EncoderAndDecoderTest ‑ test
org.apache.uniffle.common.netty.TransportFrameDecoderTest ‑ testShouldRpcRequestsToBeReleased
org.apache.uniffle.common.netty.TransportFrameDecoderTest ‑ testShouldRpcResponsesToBeReleased
org.apache.uniffle.common.netty.client.TransportClientFactoryTest ‑ testClientDiffPartition
org.apache.uniffle.common.netty.client.TransportClientFactoryTest ‑ testClientDiffServer
org.apache.uniffle.common.netty.client.TransportClientFactoryTest ‑ testClientReuse
org.apache.uniffle.common.netty.client.TransportClientFactoryTest ‑ testCreateClient
org.apache.uniffle.common.netty.protocol.NettyProtocolTest ‑ testGetLocalShuffleDataRequest
org.apache.uniffle.common.netty.protocol.NettyProtocolTest ‑ testGetLocalShuffleDataResponse
org.apache.uniffle.common.netty.protocol.NettyProtocolTest ‑ testGetLocalShuffleIndexRequest
org.apache.uniffle.common.netty.protocol.NettyProtocolTest ‑ testGetLocalShuffleIndexResponse
org.apache.uniffle.common.netty.protocol.NettyProtocolTest ‑ testGetMemoryShuffleDataRequest
org.apache.uniffle.common.netty.protocol.NettyProtocolTest ‑ testGetMemoryShuffleDataResponse
org.apache.uniffle.common.netty.protocol.NettyProtocolTest ‑ testRpcResponse
org.apache.uniffle.common.netty.protocol.NettyProtocolTest ‑ testSendShuffleDataRequest
org.apache.uniffle.common.records.RecordsReaderWriterTest ‑ testWriteAndReadRecordFile1{String, File}[1]
org.apache.uniffle.common.records.RecordsReaderWriterTest ‑ testWriteAndReadRecordFile1{String, File}[2]
org.apache.uniffle.common.records.RecordsReaderWriterTest ‑ testWriteAndReadRecordFile2{String, File}[1]
org.apache.uniffle.common.records.RecordsReaderWriterTest ‑ testWriteAndReadRecordFile2{String, File}[2]
org.apache.uniffle.common.records.RecordsReaderWriterTest ‑ testWriteAndReadRecordFile3{String, File}[1]
org.apache.uniffle.common.records.RecordsReaderWriterTest ‑ testWriteAndReadRecordFile3{String, File}[2]
org.apache.uniffle.common.records.RecordsReaderWriterTest ‑ testWriteAndReadRecordFile4{String, File}[1]
org.apache.uniffle.common.records.RecordsReaderWriterTest ‑ testWriteAndReadRecordFile4{String, File}[2]
org.apache.uniffle.common.rpc.GrpcServerTest ‑ testGrpcExecutorPool
org.apache.uniffle.common.rpc.GrpcServerTest ‑ testRandomPort
org.apache.uniffle.common.rpc.StatusCodeTest ‑ test
org.apache.uniffle.common.security.HadoopSecurityContextTest ‑ testCreateIllegalContext
org.apache.uniffle.common.security.HadoopSecurityContextTest ‑ testSecuredCallable
org.apache.uniffle.common.security.HadoopSecurityContextTest ‑ testSecuredDisableProxyUser
org.apache.uniffle.common.security.HadoopSecurityContextTest ‑ testWithOutKrb5Conf
org.apache.uniffle.common.security.SecurityContextFactoryTest ‑ testCreateHadoopSecurityContext
org.apache.uniffle.common.security.SecurityContextFactoryTest ‑ testDefaultSecurityContext
org.apache.uniffle.common.segment.FixedSizeSegmentSplitterTest ‑ testAvoidEOFException{int}[1]
org.apache.uniffle.common.segment.FixedSizeSegmentSplitterTest ‑ testAvoidEOFException{int}[2]
org.apache.uniffle.common.segment.FixedSizeSegmentSplitterTest ‑ testAvoidEOFException{int}[3]
org.apache.uniffle.common.segment.FixedSizeSegmentSplitterTest ‑ testSplit
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testConsistentWithFixSizeSplitterWhenNoSkew{int}[1]
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testConsistentWithFixSizeSplitterWhenNoSkew{int}[2]
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testConsistentWithFixSizeSplitterWhenNoSkew{int}[3]
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testConsistentWithFixSizeSplitterWhenNoSkew{int}[4]
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testConsistentWithFixSizeSplitterWhenNoSkew{int}[5]
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testConsistentWithFixSizeSplitterWhenNoSkew{int}[6]
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testDiscontinuousMapTaskIds
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testSplit
org.apache.uniffle.common.segment.LocalOrderSegmentSplitterTest ‑ testSplitForMergeContinuousSegments
org.apache.uniffle.common.serializer.PartialInputStreamTest ‑ testReadFileInputStream
org.apache.uniffle.common.serializer.PartialInputStreamTest ‑ testReadMemroyInputStream
org.apache.uniffle.common.serializer.PartialInputStreamTest ‑ testReadNullBytes
org.apache.uniffle.common.serializer.SerializerFactoryTest ‑ testGetSerializer
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeKeyValues1{String, File}[1]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeKeyValues1{String, File}[2]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeKeyValues2{String, File}[1]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeKeyValues2{String, File}[2]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeKeyValues3{String, File}[1]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeKeyValues3{String, File}[2]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeKeyValues4{String, File}[1]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeKeyValues4{String, File}[2]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeObject{Class}[1]
org.apache.uniffle.common.serializer.WritableSerializerTest ‑ testSerDeObject{Class}[2]
org.apache.uniffle.common.storage.StorageInfoUtilsTest ‑ testFromProto
org.apache.uniffle.common.storage.StorageInfoUtilsTest ‑ testToProto
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testEquals
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testFromLengths
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testFromLengthsErrors
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testLayoutGetBlockId{BlockIdLayout}[1]
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testLayoutGetBlockId{BlockIdLayout}[2]
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testLayoutGetBlockId{BlockIdLayout}[3]
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testLayoutGetBlockId{BlockIdLayout}[4]
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testLayoutGetBlockId{BlockIdLayout}[5]
org.apache.uniffle.common.util.BlockIdLayoutTest ‑ testLayoutGetBlockId{BlockIdLayout}[6]
org.apache.uniffle.common.util.BlockIdTest ‑ testEquals
org.apache.uniffle.common.util.BlockIdTest ‑ testToString
org.apache.uniffle.common.util.ByteBufUtilsTest ‑ test
org.apache.uniffle.common.util.ChecksumUtilsTest ‑ crc32ByteBufferTest
org.apache.uniffle.common.util.ChecksumUtilsTest ‑ crc32TestWithByte
org.apache.uniffle.common.util.ChecksumUtilsTest ‑ crc32TestWithByteBuff
org.apache.uniffle.common.util.ExitUtilsTest ‑ test
org.apache.uniffle.common.util.ExpiringCloseableSupplierTest ‑ stressingTestManySuppliers
org.apache.uniffle.common.util.ExpiringCloseableSupplierTest ‑ testAutoCloseable
org.apache.uniffle.common.util.ExpiringCloseableSupplierTest ‑ testCacheable
org.apache.uniffle.common.util.ExpiringCloseableSupplierTest ‑ testDelegateExtendClose
org.apache.uniffle.common.util.ExpiringCloseableSupplierTest ‑ testMultipleSupplierShouldNotInterfere
org.apache.uniffle.common.util.ExpiringCloseableSupplierTest ‑ testReClose
org.apache.uniffle.common.util.ExpiringCloseableSupplierTest ‑ testRenew
org.apache.uniffle.common.util.ExpiringCloseableSupplierTest ‑ testSerialization
org.apache.uniffle.common.util.JavaUtilsTest ‑ test
org.apache.uniffle.common.util.NettyUtilsTest ‑ test
org.apache.uniffle.common.util.RetryUtilsTest ‑ testRetry
org.apache.uniffle.common.util.RetryUtilsTest ‑ testRetryWithCondition
org.apache.uniffle.common.util.RssUtilsTest ‑ getMetricNameForHostNameTest
org.apache.uniffle.common.util.RssUtilsTest ‑ testCloneBitmap
org.apache.uniffle.common.util.RssUtilsTest ‑ testGenerateServerToPartitions
org.apache.uniffle.common.util.RssUtilsTest ‑ testGetConfiguredLocalDirs
org.apache.uniffle.common.util.RssUtilsTest ‑ testGetHostIp
org.apache.uniffle.common.util.RssUtilsTest ‑ testGetPropertiesFromFile
org.apache.uniffle.common.util.RssUtilsTest ‑ testLoadExtentions
org.apache.uniffle.common.util.RssUtilsTest ‑ testSerializeBitmap
org.apache.uniffle.common.util.RssUtilsTest ‑ testShuffleBitmapToPartitionBitmap{BlockIdLayout}[1]
org.apache.uniffle.common.util.RssUtilsTest ‑ testShuffleBitmapToPartitionBitmap{BlockIdLayout}[2]
org.apache.uniffle.common.util.RssUtilsTest ‑ testStartServiceOnPort
org.apache.uniffle.common.util.ThreadUtilsTest ‑ invokeAllTimeoutThreadPoolTest
org.apache.uniffle.common.util.ThreadUtilsTest ‑ shutdownThreadPoolTest
org.apache.uniffle.common.util.ThreadUtilsTest ‑ testExecuteTasksWithFutureHandler
org.apache.uniffle.common.util.ThreadUtilsTest ‑ testExecuteTasksWithFutureHandlerAndTimeout
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[10]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[11]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[12]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[13]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[14]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[15]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[16]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[17]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[18]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[19]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[1]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[20]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[21]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[22]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[23]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[24]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[25]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[26]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[27]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[28]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[2]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[3]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[4]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[5]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[6]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[7]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[8]
org.apache.uniffle.common.util.UnitConverterTest ‑ testByteString{Long, String, ByteUnit}[9]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[10]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[11]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[12]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[13]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[14]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[15]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[16]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[17]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[18]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[1]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[2]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[3]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[4]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[5]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[6]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[7]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[8]
org.apache.uniffle.common.util.UnitConverterTest ‑ testTimeString{Long, String, TimeUnit}[9]
org.apache.uniffle.common.web.JettyServerTest ‑ jettyServerStartTest
org.apache.uniffle.common.web.JettyServerTest ‑ jettyServerTest
org.apache.uniffle.coordinator.ApplicationManagerTest ‑ clearWithoutRemoteStorageTest
org.apache.uniffle.coordinator.ApplicationManagerTest ‑ refreshTest
org.apache.uniffle.coordinator.CoordinatorConfTest ‑ test
org.apache.uniffle.coordinator.CoordinatorServerTest ‑ test
org.apache.uniffle.coordinator.QuotaManagerTest ‑ testCheckQuota
org.apache.uniffle.coordinator.QuotaManagerTest ‑ testCheckQuotaMetrics
org.apache.uniffle.coordinator.QuotaManagerTest ‑ testCheckQuotaWithDefault
org.apache.uniffle.coordinator.QuotaManagerTest ‑ testDetectUserResource
org.apache.uniffle.coordinator.QuotaManagerTest ‑ testQuotaManagerWithoutAccessQuotaChecker
org.apache.uniffle.coordinator.ServerNodeTest ‑ compareTest
org.apache.uniffle.coordinator.ServerNodeTest ‑ testNettyPort
org.apache.uniffle.coordinator.ServerNodeTest ‑ testStorageInfoOfServerNode
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ excludeNodesNoDelayTest
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ getLostServerListTest
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ getServerListForNettyTest
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ getServerListTest
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ getUnhealthyServerList
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ heartbeatTimeoutTest
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ startupSilentPeriodTest
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ testGetCorrectServerNodesWhenOneNodeRemoved
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ testGetCorrectServerNodesWhenOneNodeRemovedAndUnhealthyNodeFound
org.apache.uniffle.coordinator.SimpleClusterManagerTest ‑ updateExcludeNodesTest
org.apache.uniffle.coordinator.access.AccessManagerTest ‑ test
org.apache.uniffle.coordinator.checker.AccessCandidatesCheckerTest ‑ test{File}
org.apache.uniffle.coordinator.checker.AccessClusterLoadCheckerTest ‑ testAccessInfoRequiredShuffleServers
org.apache.uniffle.coordinator.checker.AccessClusterLoadCheckerTest ‑ testWhenAvailableServerThresholdSpecified
org.apache.uniffle.coordinator.checker.AccessQuotaCheckerTest ‑ testAccessInfoRequiredShuffleServers
org.apache.uniffle.coordinator.conf.DynamicClientConfServiceTest ‑ testByLegacyParser{File}
org.apache.uniffle.coordinator.conf.LegacyClientConfParserTest ‑ testParse
org.apache.uniffle.coordinator.conf.RssClientConfApplyManagerTest ‑ testBypassApply
org.apache.uniffle.coordinator.conf.RssClientConfApplyManagerTest ‑ testCustomizeApplyStrategy
org.apache.uniffle.coordinator.conf.YamlClientConfParserTest ‑ testFromFile
org.apache.uniffle.coordinator.conf.YamlClientConfParserTest ‑ testParse
org.apache.uniffle.coordinator.metric.CoordinatorMetricsTest ‑ testAllMetrics
org.apache.uniffle.coordinator.metric.CoordinatorMetricsTest ‑ testCoordinatorMetrics
org.apache.uniffle.coordinator.metric.CoordinatorMetricsTest ‑ testCoordinatorMetricsWithNames
org.apache.uniffle.coordinator.metric.CoordinatorMetricsTest ‑ testDynamicMetrics
org.apache.uniffle.coordinator.metric.CoordinatorMetricsTest ‑ testGrpcMetrics
org.apache.uniffle.coordinator.metric.CoordinatorMetricsTest ‑ testJvmMetrics
org.apache.uniffle.coordinator.strategy.assignment.BasicAssignmentStrategyTest ‑ testAssign
org.apache.uniffle.coordinator.strategy.assignment.BasicAssignmentStrategyTest ‑ testAssignWithDifferentNodeNum
org.apache.uniffle.coordinator.strategy.assignment.BasicAssignmentStrategyTest ‑ testAssignmentShuffleNodesNum
org.apache.uniffle.coordinator.strategy.assignment.BasicAssignmentStrategyTest ‑ testRandomAssign
org.apache.uniffle.coordinator.strategy.assignment.BasicAssignmentStrategyTest ‑ testWithContinuousSelectPartitionStrategy
org.apache.uniffle.coordinator.strategy.assignment.PartitionBalanceAssignmentStrategyTest ‑ testAssign
org.apache.uniffle.coordinator.strategy.assignment.PartitionBalanceAssignmentStrategyTest ‑ testAssignmentShuffleNodesNum
org.apache.uniffle.coordinator.strategy.assignment.PartitionBalanceAssignmentStrategyTest ‑ testAssignmentWithMustDiff
org.apache.uniffle.coordinator.strategy.assignment.PartitionBalanceAssignmentStrategyTest ‑ testAssignmentWithNone
org.apache.uniffle.coordinator.strategy.assignment.PartitionBalanceAssignmentStrategyTest ‑ testAssignmentWithPreferDiff
org.apache.uniffle.coordinator.strategy.assignment.PartitionBalanceAssignmentStrategyTest ‑ testWithContinuousSelectPartitionStrategy
org.apache.uniffle.coordinator.strategy.assignment.PartitionRangeAssignmentTest ‑ test
org.apache.uniffle.coordinator.strategy.assignment.PartitionRangeTest ‑ test
org.apache.uniffle.coordinator.strategy.partition.ContinuousSelectPartitionStrategyTest ‑ test
org.apache.uniffle.coordinator.strategy.storage.AppBalanceSelectStorageStrategyTest ‑ selectStorageTest
org.apache.uniffle.coordinator.strategy.storage.AppBalanceSelectStorageStrategyTest ‑ storageCounterMulThreadTest
org.apache.uniffle.coordinator.strategy.storage.LowestIOSampleCostSelectStorageStrategyTest ‑ selectStorageMulThreadTest
org.apache.uniffle.coordinator.strategy.storage.LowestIOSampleCostSelectStorageStrategyTest ‑ selectStorageTest
org.apache.uniffle.coordinator.util.CoordinatorUtilsTest ‑ testExtractClusterConf
org.apache.uniffle.coordinator.util.CoordinatorUtilsTest ‑ testGenerateRanges
org.apache.uniffle.coordinator.util.CoordinatorUtilsTest ‑ testGenerateRangesGroup
org.apache.uniffle.coordinator.util.CoordinatorUtilsTest ‑ testNextId
org.apache.uniffle.coordinator.web.UniffleServicesRESTTest ‑ testGetApplications
org.apache.uniffle.coordinator.web.UniffleServicesRESTTest ‑ testGetApplicationsPage
org.apache.uniffle.coordinator.web.UniffleServicesRESTTest ‑ testGetApplicationsWithAppRegex
org.apache.uniffle.coordinator.web.UniffleServicesRESTTest ‑ testGetApplicationsWithNoFilter
org.apache.uniffle.coordinator.web.UniffleServicesRESTTest ‑ testGetApplicationsWithNull
org.apache.uniffle.coordinator.web.UniffleServicesRESTTest ‑ testGetApplicationsWithStartTimeAndEndTime
org.apache.uniffle.dashboard.web.utils.DashboardUtilsTest ‑ testConvertToMap
org.apache.uniffle.server.HealthScriptCheckerTest ‑ checkIsHealthy
org.apache.uniffle.server.KerberizedShuffleTaskManagerTest ‑ removeShuffleDataWithHdfsTest
org.apache.uniffle.server.LocalSingleStorageTypeFromEnvProviderTest ‑ testJsonSourceParse
org.apache.uniffle.server.LocalSingleStorageTypeFromEnvProviderTest ‑ testMultipleMountPoints
org.apache.uniffle.server.LocalStorageCheckerTest ‑ testCheckingStorageHang{File}
org.apache.uniffle.server.LocalStorageCheckerTest ‑ testGetUniffleUsedSpace{File}
org.apache.uniffle.server.ShuffleFlushManagerOnKerberizedHadoopTest ‑ clearTest
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ clearLocalTest{File}
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ clearTest
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ complexWriteTest
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ concurrentWrite2HdfsWriteOfSinglePartition
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ concurrentWrite2HdfsWriteOneByOne
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ defaultFlushEventHandlerTest{File}
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ fallbackWrittenWhenHybridStorageManagerEnableTest{File}
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ hadoopConfTest
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ localMetricsTest{File}
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ testCreateWriteHandlerFailed{File}
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ totalLocalFileWriteDataMetricTest
org.apache.uniffle.server.ShuffleFlushManagerTest ‑ writeTest
org.apache.uniffle.server.ShuffleServerConfTest ‑ confByStringTest
org.apache.uniffle.server.ShuffleServerConfTest ‑ confTest
org.apache.uniffle.server.ShuffleServerConfTest ‑ defaultConfTest
org.apache.uniffle.server.ShuffleServerConfTest ‑ envConfTest
org.apache.uniffle.server.ShuffleServerGrpcMetricsTest ‑ testLatencyMetrics
org.apache.uniffle.server.ShuffleServerMetricsTest ‑ testGrpcMetrics
org.apache.uniffle.server.ShuffleServerMetricsTest ‑ testHadoopStorageWriteDataSize
org.apache.uniffle.server.ShuffleServerMetricsTest ‑ testJvmMetrics
org.apache.uniffle.server.ShuffleServerMetricsTest ‑ testNettyMetrics
org.apache.uniffle.server.ShuffleServerMetricsTest ‑ testServerMetrics
org.apache.uniffle.server.ShuffleServerMetricsTest ‑ testServerMetricsConcurrently
org.apache.uniffle.server.ShuffleServerMetricsTest ‑ testStorageCounter
org.apache.uniffle.server.ShuffleServerTest ‑ decommissionTest{boolean}[1]
org.apache.uniffle.server.ShuffleServerTest ‑ decommissionTest{boolean}[2]
org.apache.uniffle.server.ShuffleServerTest ‑ nettyServerTest
org.apache.uniffle.server.ShuffleServerTest ‑ startTest
org.apache.uniffle.server.ShuffleTaskInfoTest ‑ hugePartitionConcurrentTest
org.apache.uniffle.server.ShuffleTaskInfoTest ‑ hugePartitionTest
org.apache.uniffle.server.ShuffleTaskInfoTest ‑ isHugePartitionTest
org.apache.uniffle.server.ShuffleTaskInfoTest ‑ partitionSizeSummaryTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ appPurgeWithLocalfileTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ checkAndClearLeakShuffleDataTest{File}
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ clearMultiTimesTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ clearTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ getBlockIdsByMultiPartitionTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ getBlockIdsByPartitionIdTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ hugePartitionMemoryUsageLimitTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ partitionDataSizeSummaryTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ registerShuffleTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ removeResourcesByShuffleIdsMultiTimesTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ removeShuffleDataWithHdfsTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ removeShuffleDataWithLocalfileTest
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ testAddFinishedBlockIdsWithoutRegister
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ testGetFinishedBlockIds
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ testGetMaxConcurrencyWriting
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ testRegisterShuffleAfterAppIsExpired
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ testStorageRemoveResourceHang{File}
org.apache.uniffle.server.ShuffleTaskManagerTest ‑ writeProcessTest
org.apache.uniffle.server.StorageCheckerTest ‑ checkTest{File}
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ blockSizeMetricsTest
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ bufferManagerInitTest
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ bufferSizeTest
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ cacheShuffleDataTest
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ cacheShuffleDataWithPreAllocationTest
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ flushBufferTestWhenNotSelectedStorage{File}
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ flushSingleBufferForHugePartitionTest{File}
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ flushSingleBufferTest{File}
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ getShuffleDataTest
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ getShuffleDataWithExpectedTaskIdsTest
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ registerBufferTest
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ shuffleFlushThreshold
org.apache.uniffle.server.buffer.ShuffleBufferManagerTest ‑ shuffleIdToSizeTest
org.apache.uniffle.server.buffer.ShuffleBufferWithLinkedListTest ‑ appendMultiBlocksTest
org.apache.uniffle.server.buffer.ShuffleBufferWithLinkedListTest ‑ appendTest
org.apache.uniffle.server.buffer.ShuffleBufferWithLinkedListTest ‑ getShuffleDataTest
org.apache.uniffle.server.buffer.ShuffleBufferWithLinkedListTest ‑ getShuffleDataWithExpectedTaskIdsFilterTest
org.apache.uniffle.server.buffer.ShuffleBufferWithLinkedListTest ‑ getShuffleDataWithLocalOrderTest
org.apache.uniffle.server.buffer.ShuffleBufferWithLinkedListTest ‑ toFlushEventTest
org.apache.uniffle.server.buffer.ShuffleBufferWithSkipListTest ‑ appendMultiBlocksTest
org.apache.uniffle.server.buffer.ShuffleBufferWithSkipListTest ‑ appendTest
org.apache.uniffle.server.buffer.ShuffleBufferWithSkipListTest ‑ getShuffleDataWithExpectedTaskIdsFilterTest
org.apache.uniffle.server.buffer.ShuffleBufferWithSkipListTest ‑ toFlushEventTest
org.apache.uniffle.server.merge.BlockFlushFileReaderTest ‑ writeTestWithMergeWhenInterrupted{String, File}[1]
org.apache.uniffle.server.merge.BlockFlushFileReaderTest ‑ writeTestWithMerge{String, File}[1]
org.apache.uniffle.server.merge.BlockFlushFileReaderTest ‑ writeTestWithMerge{String, File}[2]
Check notice on line 0 in .github
github-actions / Test Results
1019 tests found (test 734 to 1019)
There are 1019 tests, see "Raw output" for the list of tests 734 to 1019.
Raw output
org.apache.uniffle.server.merge.BlockFlushFileReaderTest ‑ writeTestWithMerge{String, File}[3]
org.apache.uniffle.server.merge.MergedResultTest ‑ testMergeSegmentToMergeResult{String, File}[1]
org.apache.uniffle.server.merge.MergedResultTest ‑ testMergedResult
org.apache.uniffle.server.merge.ShuffleMergeManagerTest ‑ testMergerManager{String, File}[1]
org.apache.uniffle.server.storage.HadoopStorageManagerTest ‑ testRegisterRemoteStorage
org.apache.uniffle.server.storage.HadoopStorageManagerTest ‑ testRemoveExpiredResourcesWithOneReplica{File}
org.apache.uniffle.server.storage.HadoopStorageManagerTest ‑ testRemoveExpiredResourcesWithTwoReplicas{File}
org.apache.uniffle.server.storage.HadoopStorageManagerTest ‑ testRemoveResources
org.apache.uniffle.server.storage.HybridStorageManagerTest ‑ fallbackTestWhenLocalStorageCorrupted
org.apache.uniffle.server.storage.HybridStorageManagerTest ‑ selectStorageManagerTest
org.apache.uniffle.server.storage.HybridStorageManagerTest ‑ testStorageManagerSelectorOfPreferCold
org.apache.uniffle.server.storage.HybridStorageManagerTest ‑ underStorageManagerSelectionTest
org.apache.uniffle.server.storage.LocalStorageManagerTest ‑ testEnvStorageTypeProvider
org.apache.uniffle.server.storage.LocalStorageManagerTest ‑ testGetLocalStorageInfo
org.apache.uniffle.server.storage.LocalStorageManagerTest ‑ testInitLocalStorageManager
org.apache.uniffle.server.storage.LocalStorageManagerTest ‑ testInitializeLocalStorage
org.apache.uniffle.server.storage.LocalStorageManagerTest ‑ testNewAppWhileCheckLeak{ExtensionContext}
org.apache.uniffle.server.storage.LocalStorageManagerTest ‑ testStorageSelection
org.apache.uniffle.server.storage.LocalStorageManagerTest ‑ testStorageSelectionWhenReachingHighWatermark
org.apache.uniffle.server.storage.StorageManagerFallbackStrategyTest ‑ testDefaultFallbackStrategy
org.apache.uniffle.server.storage.StorageManagerFallbackStrategyTest ‑ testHadoopFallbackStrategy
org.apache.uniffle.server.storage.StorageManagerFallbackStrategyTest ‑ testLocalFallbackStrategy
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutInsufficientConfigException{Integer, Integer, Integer, String}[1]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutInsufficientConfigException{Integer, Integer, Integer, String}[2]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutInsufficientConfigException{Integer, Integer, Integer, String}[3]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutInsufficientConfigException{Integer, Integer, Integer, String}[4]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutInsufficientConfigException{Integer, Integer, Integer, String}[5]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutInsufficientConfigException{Integer, Integer, Integer, String}[6]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutMaxPartitionsValueException{String, int, boolean}[1]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutMaxPartitionsValueException{String, int, boolean}[2]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutMaxPartitionsValueException{String, int, boolean}[3]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutMaxPartitionsValueException{String, int, boolean}[4]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutMaxPartitionsValueException{String, int, boolean}[5]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutOverrides
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutUnsupportedMaxPartitions{String, int, boolean, String}[1]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutUnsupportedMaxPartitions{String, int, boolean, String}[2]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutUnsupportedMaxPartitions{String, int, boolean, String}[3]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutUnsupportedMaxPartitions{String, int, boolean, String}[4]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutUnsupportedMaxPartitions{String, int, boolean, String}[5]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayoutUnsupportedMaxPartitions{String, int, boolean, String}[6]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[10]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[11]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[12]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[13]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[14]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[15]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[16]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[17]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[18]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[19]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[1]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[20]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[21]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[22]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[23]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[24]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[25]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[26]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[27]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[28]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[29]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[2]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[30]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[31]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[3]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[4]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[5]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[6]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[7]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[8]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testConfigureBlockIdLayout{String, Integer, Boolean, String, int, int, int}[9]
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testFetchAndApplyDynamicConf
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testGetDefaultRemoteStorageInfo
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testGetTaskAttemptIdWithSpeculation
org.apache.uniffle.shuffle.manager.RssShuffleManagerBaseTest ‑ testGetTaskAttemptIdWithoutSpeculation
org.apache.uniffle.shuffle.manager.ShuffleManagerGrpcServiceTest ‑ testShuffleManagerGrpcService
org.apache.uniffle.shuffle.manager.ShuffleManagerServerFactoryTest ‑ testShuffleManagerServerType{ServerType}[1]
org.apache.uniffle.shuffle.manager.ShuffleManagerServerFactoryTest ‑ testShuffleManagerServerType{ServerType}[2]
org.apache.uniffle.storage.common.DefaultStorageMediaProviderTest ‑ getGetDeviceName
org.apache.uniffle.storage.common.DefaultStorageMediaProviderTest ‑ getGetFileStore{File}
org.apache.uniffle.storage.common.DefaultStorageMediaProviderTest ‑ testStorageProvider
org.apache.uniffle.storage.common.LocalStorageTest ‑ baseDirectoryInitTest
org.apache.uniffle.storage.common.LocalStorageTest ‑ canWriteTest
org.apache.uniffle.storage.common.LocalStorageTest ‑ canWriteTestWithDiskCapacityCheck
org.apache.uniffle.storage.common.LocalStorageTest ‑ diskStorageInfoTest
org.apache.uniffle.storage.common.LocalStorageTest ‑ getCapacityInitTest
org.apache.uniffle.storage.common.LocalStorageTest ‑ writeHandlerTest
org.apache.uniffle.storage.common.ShuffleFileInfoTest ‑ test
org.apache.uniffle.storage.handler.impl.HadoopClientReadHandlerTest ‑ test
org.apache.uniffle.storage.handler.impl.HadoopFileReaderTest ‑ createStreamAppendTest
org.apache.uniffle.storage.handler.impl.HadoopFileReaderTest ‑ createStreamTest
org.apache.uniffle.storage.handler.impl.HadoopFileReaderTest ‑ readDataTest
org.apache.uniffle.storage.handler.impl.HadoopFileWriterTest ‑ createStreamAppendTest
org.apache.uniffle.storage.handler.impl.HadoopFileWriterTest ‑ createStreamDirectory
org.apache.uniffle.storage.handler.impl.HadoopFileWriterTest ‑ createStreamFirstTest
org.apache.uniffle.storage.handler.impl.HadoopFileWriterTest ‑ createStreamTest
org.apache.uniffle.storage.handler.impl.HadoopFileWriterTest ‑ writeBufferArrayTest
org.apache.uniffle.storage.handler.impl.HadoopFileWriterTest ‑ writeBufferTest
org.apache.uniffle.storage.handler.impl.HadoopFileWriterTest ‑ writeSegmentTest
org.apache.uniffle.storage.handler.impl.HadoopHandlerTest ‑ initTest
org.apache.uniffle.storage.handler.impl.HadoopHandlerTest ‑ writeTest
org.apache.uniffle.storage.handler.impl.HadoopShuffleReadHandlerTest ‑ test
org.apache.uniffle.storage.handler.impl.HadoopShuffleReadHandlerTest ‑ testDataInconsistent
org.apache.uniffle.storage.handler.impl.KerberizedHadoopClientReadHandlerTest ‑ test
org.apache.uniffle.storage.handler.impl.KerberizedHadoopShuffleReadHandlerTest ‑ test
org.apache.uniffle.storage.handler.impl.LocalFileHandlerTest ‑ testReadIndex
org.apache.uniffle.storage.handler.impl.LocalFileHandlerTest ‑ writeBigDataTest{File}
org.apache.uniffle.storage.handler.impl.LocalFileHandlerTest ‑ writeTest{File}
org.apache.uniffle.storage.handler.impl.LocalFileServerReadHandlerTest ‑ testDataInconsistent
org.apache.uniffle.storage.handler.impl.PooledHadoopShuffleWriteHandlerTest ‑ concurrentWrite
org.apache.uniffle.storage.handler.impl.PooledHadoopShuffleWriteHandlerTest ‑ lazyInitializeWriterHandlerTest
org.apache.uniffle.storage.handler.impl.PooledHadoopShuffleWriteHandlerTest ‑ writeSameFileWhenNoRaceCondition
org.apache.uniffle.storage.util.ShuffleHadoopStorageUtilsTest ‑ testUploadFile{File}
org.apache.uniffle.storage.util.ShuffleKerberizedHadoopStorageUtilsTest ‑ testUploadFile{File}
org.apache.uniffle.storage.util.ShuffleStorageUtilsTest ‑ getPartitionRangeTest
org.apache.uniffle.storage.util.ShuffleStorageUtilsTest ‑ getShuffleDataPathWithRangeTest
org.apache.uniffle.storage.util.ShuffleStorageUtilsTest ‑ getStorageIndexTest
org.apache.uniffle.storage.util.ShuffleStorageUtilsTest ‑ mergeSegmentsTest
org.apache.uniffle.storage.util.StorageTypeTest ‑ commonTest
org.apache.uniffle.test.AQERepartitionTest ‑ resultCompareTest
org.apache.uniffle.test.AQESkewedJoinTest ‑ resultCompareTest
org.apache.uniffle.test.AQESkewedJoinWithLocalOrderTest ‑ resultCompareTest
org.apache.uniffle.test.AccessCandidatesCheckerHadoopTest ‑ test
org.apache.uniffle.test.AccessCandidatesCheckerKerberizedHadoopTest ‑ test
org.apache.uniffle.test.AccessClusterTest ‑ testUsingCustomExtraProperties
org.apache.uniffle.test.AccessClusterTest ‑ test{File}
org.apache.uniffle.test.AssignmentWithTagsTest ‑ testTags
org.apache.uniffle.test.AutoAccessTest ‑ test
org.apache.uniffle.test.CombineByKeyTest ‑ combineByKeyTest
org.apache.uniffle.test.ContinuousSelectPartitionStrategyTest ‑ resultCompareTest
org.apache.uniffle.test.CoordinatorAdminServiceTest ‑ test
org.apache.uniffle.test.CoordinatorAssignmentTest ‑ testAssignmentServerNodesNumber
org.apache.uniffle.test.CoordinatorAssignmentTest ‑ testGetReShuffleAssignments
org.apache.uniffle.test.CoordinatorAssignmentTest ‑ testSilentPeriod
org.apache.uniffle.test.CoordinatorGrpcServerTest ‑ testGrpcConnectionSize
org.apache.uniffle.test.CoordinatorGrpcTest ‑ appHeartbeatTest
org.apache.uniffle.test.CoordinatorGrpcTest ‑ getShuffleAssignmentsTest
org.apache.uniffle.test.CoordinatorGrpcTest ‑ getShuffleRegisterInfoTest
org.apache.uniffle.test.CoordinatorGrpcTest ‑ rpcMetricsTest
org.apache.uniffle.test.CoordinatorGrpcTest ‑ shuffleServerHeartbeatTest
org.apache.uniffle.test.CoordinatorGrpcTest ‑ testGetPartitionToServers
org.apache.uniffle.test.CoordinatorReconfigureNodeMaxTest ‑ testReconfigureNodeMax
org.apache.uniffle.test.DynamicClientConfServiceHadoopTest ‑ test
org.apache.uniffle.test.DynamicClientConfServiceKerberlizedHadoopTest ‑ testConfInHadoop
org.apache.uniffle.test.DynamicConfTest ‑ dynamicConfTest
org.apache.uniffle.test.DynamicFetchClientConfTest ‑ test
org.apache.uniffle.test.FailingTasksTest ‑ testFailedTasks
org.apache.uniffle.test.FetchClientConfTest ‑ testFetchRemoteStorageByApp{File}
org.apache.uniffle.test.FetchClientConfTest ‑ testFetchRemoteStorageByIO{File}
org.apache.uniffle.test.FetchClientConfTest ‑ test{File}
org.apache.uniffle.test.GetReaderTest ‑ test
org.apache.uniffle.test.GetShuffleReportForMultiPartTest ‑ resultCompareTest
org.apache.uniffle.test.GroupByKeyTest ‑ groupByTest
org.apache.uniffle.test.HadoopConfTest ‑ hadoopConfTest
org.apache.uniffle.test.HealthCheckCoordinatorGrpcTest ‑ healthCheckTest
org.apache.uniffle.test.HealthCheckTest ‑ buildInCheckerTest
org.apache.uniffle.test.HealthCheckTest ‑ checkTest
org.apache.uniffle.test.LargeSorterTest ‑ largeSorterTest
org.apache.uniffle.test.MapSideCombineTest ‑ resultCompareTest
org.apache.uniffle.test.NullOfKeyOrValueTest ‑ nullOfKeyOrValueTest
org.apache.uniffle.test.PartitionBalanceCoordinatorGrpcTest ‑ getShuffleAssignmentsTest
org.apache.uniffle.test.PartitionBlockDataReassignBasicTest ‑ resultCompareTest
org.apache.uniffle.test.PartitionBlockDataReassignMultiTimesTest ‑ resultCompareTest
org.apache.uniffle.test.QuorumTest ‑ case1
org.apache.uniffle.test.QuorumTest ‑ case10
org.apache.uniffle.test.QuorumTest ‑ case11
org.apache.uniffle.test.QuorumTest ‑ case12
org.apache.uniffle.test.QuorumTest ‑ case2
org.apache.uniffle.test.QuorumTest ‑ case3
org.apache.uniffle.test.QuorumTest ‑ case4
org.apache.uniffle.test.QuorumTest ‑ case5{File}
org.apache.uniffle.test.QuorumTest ‑ case6
org.apache.uniffle.test.QuorumTest ‑ case7
org.apache.uniffle.test.QuorumTest ‑ case8
org.apache.uniffle.test.QuorumTest ‑ case9
org.apache.uniffle.test.QuorumTest ‑ quorumConfigTest
org.apache.uniffle.test.QuorumTest ‑ rpcFailedTest
org.apache.uniffle.test.RSSStageDynamicServerReWriteTest ‑ testRSSStageResubmit
org.apache.uniffle.test.RSSStageResubmitTest ‑ testRSSStageResubmit
org.apache.uniffle.test.ReassignAndStageRetryTest ‑ resultCompareTest
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTest ‑ remoteMergeWriteReadTestMultiPartitionWithCombine{String}[1]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTest ‑ remoteMergeWriteReadTestMultiPartitionWithCombine{String}[2]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTest ‑ remoteMergeWriteReadTestMultiPartition{String}[1]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTest ‑ remoteMergeWriteReadTestMultiPartition{String}[2]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTest ‑ remoteMergeWriteReadTestWithCombine{String}[1]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTest ‑ remoteMergeWriteReadTestWithCombine{String}[2]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTest ‑ remoteMergeWriteReadTest{String}[1]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTest ‑ remoteMergeWriteReadTest{String}[2]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTestWhenShuffleFlushed ‑ remoteMergeWriteReadTestMultiPartitionWithCombine{String}[1]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTestWhenShuffleFlushed ‑ remoteMergeWriteReadTestMultiPartitionWithCombine{String}[2]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTestWhenShuffleFlushed ‑ remoteMergeWriteReadTestMultiPartition{String}[1]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTestWhenShuffleFlushed ‑ remoteMergeWriteReadTestMultiPartition{String}[2]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTestWhenShuffleFlushed ‑ remoteMergeWriteReadTestWithCombine{String}[1]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTestWhenShuffleFlushed ‑ remoteMergeWriteReadTestWithCombine{String}[2]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTestWhenShuffleFlushed ‑ remoteMergeWriteReadTest{String}[1]
org.apache.uniffle.test.RemoteMergeShuffleWithRssClientTestWhenShuffleFlushed ‑ remoteMergeWriteReadTest{String}[2]
org.apache.uniffle.test.RepartitionWithHadoopHybridStorageRssTest ‑ resultCompareTest
org.apache.uniffle.test.RepartitionWithLocalFileRssTest ‑ resultCompareTest
org.apache.uniffle.test.RepartitionWithMemoryHybridStorageRssTest ‑ resultCompareTest
org.apache.uniffle.test.RepartitionWithMemoryRssTest ‑ resultCompareTest
org.apache.uniffle.test.RepartitionWithMemoryRssTest ‑ testMemoryRelease
org.apache.uniffle.test.RpcClientRetryTest ‑ testRpcRetryLogic{StorageType}[1]
org.apache.uniffle.test.RpcClientRetryTest ‑ testRpcRetryLogic{StorageType}[2]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManagerClientConfOverride{boolean}[1]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManagerClientConfOverride{boolean}[2]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManagerClientConf{BlockIdLayout}[1]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManagerClientConf{BlockIdLayout}[2]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManagerClientConf{BlockIdLayout}[3]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManagerDynamicClientConf{BlockIdLayout}[1]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManagerDynamicClientConf{BlockIdLayout}[2]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManagerDynamicClientConf{BlockIdLayout}[3]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManager{boolean}[1]
org.apache.uniffle.test.RssShuffleManagerTest ‑ testRssShuffleManager{boolean}[2]
org.apache.uniffle.test.SecondarySortTest ‑ secondarySortTest
org.apache.uniffle.test.ServletTest ‑ testDecommissionServlet
org.apache.uniffle.test.ServletTest ‑ testDecommissionSingleNode
org.apache.uniffle.test.ServletTest ‑ testDecommissionedNodeServlet
org.apache.uniffle.test.ServletTest ‑ testGetSingleNode
org.apache.uniffle.test.ServletTest ‑ testLostNodesServlet
org.apache.uniffle.test.ServletTest ‑ testNodesServlet
org.apache.uniffle.test.ServletTest ‑ testRequestWithWrongCredentials
org.apache.uniffle.test.ServletTest ‑ testUnhealthyNodesServlet
org.apache.uniffle.test.ShuffleServerConcurrentWriteOfHadoopTest ‑ testConcurrentWrite2Hadoop{int, int, boolean}[1]
org.apache.uniffle.test.ShuffleServerConcurrentWriteOfHadoopTest ‑ testConcurrentWrite2Hadoop{int, int, boolean}[2]
org.apache.uniffle.test.ShuffleServerConcurrentWriteOfHadoopTest ‑ testConcurrentWrite2Hadoop{int, int, boolean}[3]
org.apache.uniffle.test.ShuffleServerConcurrentWriteOfHadoopTest ‑ testConcurrentWrite2Hadoop{int, int, boolean}[4]
org.apache.uniffle.test.ShuffleServerGrpcTest ‑ clearResourceTest
org.apache.uniffle.test.ShuffleServerGrpcTest ‑ multipleShuffleResultTest{BlockIdLayout}[1]
org.apache.uniffle.test.ShuffleServerGrpcTest ‑ multipleShuffleResultTest{BlockIdLayout}[2]
org.apache.uniffle.test.ShuffleServerGrpcTest ‑ registerTest
org.apache.uniffle.test.ShuffleServerGrpcTest ‑ rpcMetricsTest
org.apache.uniffle.test.ShuffleServerGrpcTest ‑ sendDataWithoutRequirePreAllocation
org.apache.uniffle.test.ShuffleServerGrpcTest ‑ shuffleResultTest
org.apache.uniffle.test.ShuffleServerInternalGrpcTest ‑ decommissionTest
org.apache.uniffle.test.ShuffleServerOnRandomPortTest ‑ startGrpcServerOnRandomPort
org.apache.uniffle.test.ShuffleServerOnRandomPortTest ‑ startStreamServerOnRandomPort
org.apache.uniffle.test.ShuffleServerWithLocalOfExceptionTest ‑ testReadWhenConnectionFailedShouldThrowException
org.apache.uniffle.test.ShuffleServerWithMemLocalHadoopTest ‑ memoryLocalFileHadoopReadWithFilterTest{boolean, boolean}[1]
org.apache.uniffle.test.ShuffleServerWithMemLocalHadoopTest ‑ memoryLocalFileHadoopReadWithFilterTest{boolean, boolean}[2]
org.apache.uniffle.test.ShuffleServerWithMemLocalHadoopTest ‑ memoryLocalFileHadoopReadWithFilterTest{boolean, boolean}[3]
org.apache.uniffle.test.ShuffleServerWithMemLocalHadoopTest ‑ memoryLocalFileHadoopReadWithFilterTest{boolean, boolean}[4]
org.apache.uniffle.test.ShuffleUnregisterWithHadoopTest ‑ unregisterShuffleTest
org.apache.uniffle.test.ShuffleUnregisterWithLocalfileTest ‑ unregisterShuffleTest
org.apache.uniffle.test.ShuffleWithRssClientTest ‑ emptyTaskTest
org.apache.uniffle.test.ShuffleWithRssClientTest ‑ reportBlocksToShuffleServerIfNecessary
org.apache.uniffle.test.ShuffleWithRssClientTest ‑ reportMultipleServerTest
org.apache.uniffle.test.ShuffleWithRssClientTest ‑ rpcFailTest
org.apache.uniffle.test.ShuffleWithRssClientTest ‑ testRetryAssgin
org.apache.uniffle.test.ShuffleWithRssClientTest ‑ writeReadTest
org.apache.uniffle.test.SimpleShuffleServerManagerTest ‑ testClientAndServerConnections
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest10{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest10{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest1{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest1{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest2{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest2{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest3{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest3{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest4{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest4{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest5{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest5{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest6{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest6{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest7{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest7{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest8{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest8{boolean}[2]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest9{boolean}[1]
org.apache.uniffle.test.SparkClientWithLocalTest ‑ readTest9{boolean}[2]
org.apache.uniffle.test.SparkSQLWithDelegationShuffleManagerFallbackTest ‑ resultCompareTest
org.apache.uniffle.test.SparkSQLWithDelegationShuffleManagerTest ‑ resultCompareTest
org.apache.uniffle.test.SparkSQLWithMemoryLocalTest ‑ resultCompareTest
org.apache.uniffle.test.TezCartesianProductTest ‑ cartesianProductTest
org.apache.uniffle.test.TezHashJoinTest ‑ hashJoinDoBroadcastTest
org.apache.uniffle.test.TezHashJoinTest ‑ hashJoinTest
org.apache.uniffle.test.TezOrderedWordCountTest ‑ orderedWordCountTest
org.apache.uniffle.test.TezSimpleSessionExampleTest ‑ simpleSessionExampleTest
org.apache.uniffle.test.TezSortMergeJoinTest ‑ sortMergeJoinTest
org.apache.uniffle.test.TezWordCountTest ‑ wordCountTest
org.apache.uniffle.test.TezWordCountWithFailuresTest ‑ wordCountTestWithNodeUnhealthyWhenAvoidRecomputeDisable
org.apache.uniffle.test.TezWordCountWithFailuresTest ‑ wordCountTestWithNodeUnhealthyWhenAvoidRecomputeEnable
org.apache.uniffle.test.TezWordCountWithFailuresTest ‑ wordCountTestWithTaskFailureWhenAvoidRecomputeDisable
org.apache.uniffle.test.TezWordCountWithFailuresTest ‑ wordCountTestWithTaskFailureWhenAvoidRecomputeEnable
org.apache.uniffle.test.WordCountTest ‑ wordCountTest
org.apache.uniffle.test.WriteAndReadMetricsTest ‑ test