2024-12-02 17:24:33,104 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@30f842ca 2024-12-02 17:24:33,119 main DEBUG Took 0.012288 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-02 17:24:33,120 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-02 17:24:33,120 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-02 17:24:33,121 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-02 17:24:33,123 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,131 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-02 17:24:33,147 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,148 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,149 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,150 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,150 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,151 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,152 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,152 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,152 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,153 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,154 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,154 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,155 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,155 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,156 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,156 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,157 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,157 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,158 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,158 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,159 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,159 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,159 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,160 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 17:24:33,160 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,161 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-02 17:24:33,162 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 17:24:33,164 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-02 17:24:33,166 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-02 17:24:33,167 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-02 17:24:33,168 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-02 17:24:33,169 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-02 17:24:33,180 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-02 17:24:33,183 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-02 17:24:33,185 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-02 17:24:33,185 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-02 17:24:33,186 main DEBUG createAppenders(={Console}) 2024-12-02 17:24:33,187 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@30f842ca initialized 2024-12-02 17:24:33,187 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@30f842ca 2024-12-02 17:24:33,187 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@30f842ca OK. 2024-12-02 17:24:33,188 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-02 17:24:33,188 main DEBUG OutputStream closed 2024-12-02 17:24:33,189 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-02 17:24:33,189 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-02 17:24:33,189 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@6404f418 OK 2024-12-02 17:24:33,261 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-02 17:24:33,263 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-02 17:24:33,264 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-02 17:24:33,264 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-02 17:24:33,265 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-02 17:24:33,265 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-02 17:24:33,265 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-02 17:24:33,266 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-02 17:24:33,266 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-02 17:24:33,266 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-02 17:24:33,267 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-02 17:24:33,267 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-02 17:24:33,267 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-02 17:24:33,267 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-02 17:24:33,268 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-02 17:24:33,268 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-02 17:24:33,268 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-02 17:24:33,269 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-02 17:24:33,271 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-02 17:24:33,271 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-logging/target/hbase-logging-3.0.0-beta-2-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@6dab9b6d) with optional ClassLoader: null 2024-12-02 17:24:33,272 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-02 17:24:33,272 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@6dab9b6d] started OK. 2024-12-02T17:24:33,581 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b 2024-12-02 17:24:33,586 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-02 17:24:33,586 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-02T17:24:33,600 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestLogRolling timeout: 13 mins 2024-12-02T17:24:33,641 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testSlowSyncLogRolling Thread=12, OpenFileDescriptor=287, MaxFileDescriptor=1048576, SystemLoadAverage=419, ProcessCount=11, AvailableMemoryMB=2987 2024-12-02T17:24:33,644 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T17:24:33,660 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1, deleteOnExit=true 2024-12-02T17:24:33,661 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-02T17:24:33,662 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/test.cache.data in system properties and HBase conf 2024-12-02T17:24:33,662 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T17:24:33,663 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.log.dir in system properties and HBase conf 2024-12-02T17:24:33,664 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T17:24:33,664 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T17:24:33,664 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T17:24:33,777 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-02T17:24:33,870 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T17:24:33,874 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:24:33,875 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:24:33,875 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T17:24:33,875 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:24:33,876 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T17:24:33,876 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T17:24:33,877 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:24:33,877 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:24:33,878 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T17:24:33,878 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/nfs.dump.dir in system properties and HBase conf 2024-12-02T17:24:33,878 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/java.io.tmpdir in system properties and HBase conf 2024-12-02T17:24:33,879 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:24:33,879 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T17:24:33,880 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T17:24:34,377 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:24:34,743 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-02T17:24:34,821 INFO [Time-limited test {}] log.Log(170): Logging initialized @2492ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-02T17:24:34,899 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:24:34,968 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:24:34,991 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:24:34,991 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:24:34,993 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:24:35,009 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:24:35,012 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@ddc8467{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:24:35,013 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@70be1389{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:24:35,208 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@735fa16a{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/java.io.tmpdir/jetty-localhost-35077-hadoop-hdfs-3_4_1-tests_jar-_-any-15260750519702862108/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:24:35,218 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@6c26a5a3{HTTP/1.1, (http/1.1)}{localhost:35077} 2024-12-02T17:24:35,218 INFO [Time-limited test {}] server.Server(415): Started @2890ms 2024-12-02T17:24:35,251 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:24:35,614 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:24:35,620 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:24:35,621 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:24:35,622 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:24:35,622 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:24:35,623 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@28778f0f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:24:35,624 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@371e191c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:24:35,753 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7b07d1ba{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/java.io.tmpdir/jetty-localhost-41641-hadoop-hdfs-3_4_1-tests_jar-_-any-2839362583206694391/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:24:35,754 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@43e0a762{HTTP/1.1, (http/1.1)}{localhost:41641} 2024-12-02T17:24:35,754 INFO [Time-limited test {}] server.Server(415): Started @3426ms 2024-12-02T17:24:35,811 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:24:35,944 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:24:35,950 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:24:35,952 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:24:35,952 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:24:35,953 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:24:35,954 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@11effdcd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:24:35,954 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2d48d695{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:24:36,089 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1bf97579{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/java.io.tmpdir/jetty-localhost-33311-hadoop-hdfs-3_4_1-tests_jar-_-any-14923598461780569882/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:24:36,090 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@22b88bcb{HTTP/1.1, (http/1.1)}{localhost:33311} 2024-12-02T17:24:36,091 INFO [Time-limited test {}] server.Server(415): Started @3763ms 2024-12-02T17:24:36,094 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:24:36,288 WARN [Thread-97 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/data/data2/current/BP-25736929-172.17.0.3-1733160274475/current, will proceed with Du for space computation calculation, 2024-12-02T17:24:36,289 WARN [Thread-95 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/data/data1/current/BP-25736929-172.17.0.3-1733160274475/current, will proceed with Du for space computation calculation, 2024-12-02T17:24:36,288 WARN [Thread-98 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/data/data4/current/BP-25736929-172.17.0.3-1733160274475/current, will proceed with Du for space computation calculation, 2024-12-02T17:24:36,288 WARN [Thread-96 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/data/data3/current/BP-25736929-172.17.0.3-1733160274475/current, will proceed with Du for space computation calculation, 2024-12-02T17:24:36,348 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:24:36,361 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:24:36,441 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x37f134bb72a58682 with lease ID 0x4b72a7ceb839dd31: Processing first storage report for DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579 from datanode DatanodeRegistration(127.0.0.1:33621, datanodeUuid=52b09fe7-8d70-4f4b-88b2-9b7bb07d5e81, infoPort=34997, infoSecurePort=0, ipcPort=36907, storageInfo=lv=-57;cid=testClusterID;nsid=174267827;c=1733160274475) 2024-12-02T17:24:36,443 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x37f134bb72a58682 with lease ID 0x4b72a7ceb839dd31: from storage DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579 node DatanodeRegistration(127.0.0.1:33621, datanodeUuid=52b09fe7-8d70-4f4b-88b2-9b7bb07d5e81, infoPort=34997, infoSecurePort=0, ipcPort=36907, storageInfo=lv=-57;cid=testClusterID;nsid=174267827;c=1733160274475), blocks: 0, hasStaleStorage: true, processing time: 3 msecs, invalidatedBlocks: 0 2024-12-02T17:24:36,444 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xea21fef2b2ac12de with lease ID 0x4b72a7ceb839dd32: Processing first storage report for DS-41c49996-612d-45af-b6c8-724ff3d6f8f0 from datanode DatanodeRegistration(127.0.0.1:44647, datanodeUuid=8d44a392-54db-45ff-91f0-10718f9af9bc, infoPort=33273, infoSecurePort=0, ipcPort=43417, storageInfo=lv=-57;cid=testClusterID;nsid=174267827;c=1733160274475) 2024-12-02T17:24:36,444 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xea21fef2b2ac12de with lease ID 0x4b72a7ceb839dd32: from storage DS-41c49996-612d-45af-b6c8-724ff3d6f8f0 node DatanodeRegistration(127.0.0.1:44647, datanodeUuid=8d44a392-54db-45ff-91f0-10718f9af9bc, infoPort=33273, infoSecurePort=0, ipcPort=43417, storageInfo=lv=-57;cid=testClusterID;nsid=174267827;c=1733160274475), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:24:36,445 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xea21fef2b2ac12de with lease ID 0x4b72a7ceb839dd32: Processing first storage report for DS-f64c279f-2e60-4758-9fed-e74b30b25232 from datanode DatanodeRegistration(127.0.0.1:44647, datanodeUuid=8d44a392-54db-45ff-91f0-10718f9af9bc, infoPort=33273, infoSecurePort=0, ipcPort=43417, storageInfo=lv=-57;cid=testClusterID;nsid=174267827;c=1733160274475) 2024-12-02T17:24:36,445 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xea21fef2b2ac12de with lease ID 0x4b72a7ceb839dd32: from storage DS-f64c279f-2e60-4758-9fed-e74b30b25232 node DatanodeRegistration(127.0.0.1:44647, datanodeUuid=8d44a392-54db-45ff-91f0-10718f9af9bc, infoPort=33273, infoSecurePort=0, ipcPort=43417, storageInfo=lv=-57;cid=testClusterID;nsid=174267827;c=1733160274475), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T17:24:36,445 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x37f134bb72a58682 with lease ID 0x4b72a7ceb839dd31: Processing first storage report for DS-ef74b74c-a6b7-4859-913f-d6514889b235 from datanode DatanodeRegistration(127.0.0.1:33621, datanodeUuid=52b09fe7-8d70-4f4b-88b2-9b7bb07d5e81, infoPort=34997, infoSecurePort=0, ipcPort=36907, storageInfo=lv=-57;cid=testClusterID;nsid=174267827;c=1733160274475) 2024-12-02T17:24:36,446 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x37f134bb72a58682 with lease ID 0x4b72a7ceb839dd31: from storage DS-ef74b74c-a6b7-4859-913f-d6514889b235 node DatanodeRegistration(127.0.0.1:33621, datanodeUuid=52b09fe7-8d70-4f4b-88b2-9b7bb07d5e81, infoPort=34997, infoSecurePort=0, ipcPort=36907, storageInfo=lv=-57;cid=testClusterID;nsid=174267827;c=1733160274475), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:24:36,516 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b 2024-12-02T17:24:36,613 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/zookeeper_0, clientPort=65020, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T17:24:36,623 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=65020 2024-12-02T17:24:36,632 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:24:36,635 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:24:36,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:24:36,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:24:37,311 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0 with version=8 2024-12-02T17:24:37,311 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase-staging 2024-12-02T17:24:37,412 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-02T17:24:37,655 INFO [Time-limited test {}] client.ConnectionUtils(128): master/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:24:37,667 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:24:37,667 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:24:37,671 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:24:37,671 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:24:37,672 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:24:37,809 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T17:24:37,870 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-02T17:24:37,879 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-02T17:24:37,883 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:24:37,910 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 9921 (auto-detected) 2024-12-02T17:24:37,911 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:03 (auto-detected) 2024-12-02T17:24:37,930 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:37801 2024-12-02T17:24:37,952 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:37801 connecting to ZooKeeper ensemble=127.0.0.1:65020 2024-12-02T17:24:37,984 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:378010x0, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:24:37,988 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:37801-0x1009c051d0c0000 connected 2024-12-02T17:24:38,014 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:24:38,017 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:24:38,028 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:24:38,032 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0, hbase.cluster.distributed=false 2024-12-02T17:24:38,055 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:24:38,060 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37801 2024-12-02T17:24:38,063 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37801 2024-12-02T17:24:38,063 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37801 2024-12-02T17:24:38,071 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37801 2024-12-02T17:24:38,071 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37801 2024-12-02T17:24:38,185 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:24:38,187 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:24:38,188 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:24:38,188 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:24:38,188 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:24:38,188 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:24:38,191 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T17:24:38,193 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:24:38,194 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:38409 2024-12-02T17:24:38,196 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:38409 connecting to ZooKeeper ensemble=127.0.0.1:65020 2024-12-02T17:24:38,197 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:24:38,202 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:24:38,213 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:384090x0, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:24:38,214 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:38409-0x1009c051d0c0001 connected 2024-12-02T17:24:38,214 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:24:38,219 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T17:24:38,227 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T17:24:38,229 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T17:24:38,234 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:24:38,235 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38409 2024-12-02T17:24:38,235 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38409 2024-12-02T17:24:38,236 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38409 2024-12-02T17:24:38,238 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38409 2024-12-02T17:24:38,238 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38409 2024-12-02T17:24:38,254 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;3b3375d1a925:37801 2024-12-02T17:24:38,255 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/3b3375d1a925,37801,1733160277462 2024-12-02T17:24:38,263 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:24:38,263 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:24:38,265 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/3b3375d1a925,37801,1733160277462 2024-12-02T17:24:38,295 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:38,295 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T17:24:38,295 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:38,296 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T17:24:38,297 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/3b3375d1a925,37801,1733160277462 from backup master directory 2024-12-02T17:24:38,300 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/3b3375d1a925,37801,1733160277462 2024-12-02T17:24:38,300 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:24:38,300 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:24:38,301 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:24:38,301 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=3b3375d1a925,37801,1733160277462 2024-12-02T17:24:38,303 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-02T17:24:38,305 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-02T17:24:38,374 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase.id] with ID: ee9ea7b4-ea45-493c-bef1-e3bf24bf530b 2024-12-02T17:24:38,375 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/.tmp/hbase.id 2024-12-02T17:24:38,388 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:24:38,389 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:24:38,791 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/.tmp/hbase.id]:[hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase.id] 2024-12-02T17:24:38,838 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:24:38,843 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T17:24:38,862 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 17ms. 2024-12-02T17:24:38,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:38,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:38,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:24:38,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:24:38,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:24:38,902 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T17:24:38,908 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:24:38,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:24:38,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:24:38,958 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store 2024-12-02T17:24:38,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:24:38,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:24:38,983 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-02T17:24:38,986 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:24:38,987 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:24:38,987 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:24:38,987 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:24:38,989 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:24:38,989 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:24:38,989 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:24:38,990 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160278987Disabling compacts and flushes for region at 1733160278987Disabling writes for close at 1733160278989 (+2 ms)Writing region close event to WAL at 1733160278989Closed at 1733160278989 2024-12-02T17:24:38,992 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/.initializing 2024-12-02T17:24:38,992 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/WALs/3b3375d1a925,37801,1733160277462 2024-12-02T17:24:39,013 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C37801%2C1733160277462, suffix=, logDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/WALs/3b3375d1a925,37801,1733160277462, archiveDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/oldWALs, maxLogs=10 2024-12-02T17:24:39,022 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37801%2C1733160277462.1733160279018 2024-12-02T17:24:39,040 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/WALs/3b3375d1a925,37801,1733160277462/3b3375d1a925%2C37801%2C1733160277462.1733160279018 2024-12-02T17:24:39,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:33273:33273),(127.0.0.1/127.0.0.1:34997:34997)] 2024-12-02T17:24:39,050 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:24:39,050 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:24:39,053 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,054 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,090 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,117 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T17:24:39,120 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,123 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:39,124 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,127 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T17:24:39,127 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,128 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:24:39,128 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,131 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T17:24:39,131 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,132 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:24:39,133 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,136 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T17:24:39,136 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,137 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:24:39,137 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,140 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,141 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,146 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,147 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,150 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T17:24:39,154 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:24:39,159 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:24:39,160 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=786419, jitterRate=-1.7523765563964844E-5}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T17:24:39,166 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733160279066Initializing all the Stores at 1733160279068 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160279068Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160279069 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160279069Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160279069Cleaning up temporary data from old regions at 1733160279147 (+78 ms)Region opened successfully at 1733160279166 (+19 ms) 2024-12-02T17:24:39,167 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T17:24:39,202 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@27004af2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:24:39,235 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T17:24:39,247 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T17:24:39,248 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T17:24:39,251 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T17:24:39,253 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-02T17:24:39,258 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 5 msec 2024-12-02T17:24:39,259 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T17:24:39,288 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T17:24:39,297 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T17:24:39,300 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T17:24:39,302 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T17:24:39,304 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T17:24:39,306 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T17:24:39,308 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T17:24:39,311 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T17:24:39,313 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T17:24:39,315 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T17:24:39,316 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T17:24:39,334 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T17:24:39,336 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T17:24:39,341 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:24:39,341 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:24:39,341 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:39,341 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:39,344 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=3b3375d1a925,37801,1733160277462, sessionid=0x1009c051d0c0000, setting cluster-up flag (Was=false) 2024-12-02T17:24:39,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:39,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:39,364 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T17:24:39,366 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,37801,1733160277462 2024-12-02T17:24:39,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:39,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:39,383 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T17:24:39,387 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,37801,1733160277462 2024-12-02T17:24:39,395 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T17:24:39,468 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(746): ClusterId : ee9ea7b4-ea45-493c-bef1-e3bf24bf530b 2024-12-02T17:24:39,472 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T17:24:39,486 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T17:24:39,486 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T17:24:39,497 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T17:24:39,496 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T17:24:39,498 DEBUG [RS:0;3b3375d1a925:38409 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@772a2895, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:24:39,509 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T17:24:39,513 DEBUG [RS:0;3b3375d1a925:38409 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;3b3375d1a925:38409 2024-12-02T17:24:39,517 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T17:24:39,517 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T17:24:39,517 DEBUG [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T17:24:39,519 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T17:24:39,523 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(2659): reportForDuty to master=3b3375d1a925,37801,1733160277462 with port=38409, startcode=1733160278146 2024-12-02T17:24:39,526 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 3b3375d1a925,37801,1733160277462 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T17:24:39,534 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:24:39,534 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:24:39,535 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:24:39,535 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:24:39,535 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/3b3375d1a925:0, corePoolSize=10, maxPoolSize=10 2024-12-02T17:24:39,535 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,535 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:24:39,535 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,536 DEBUG [RS:0;3b3375d1a925:38409 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T17:24:39,541 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733160309541 2024-12-02T17:24:39,542 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:24:39,542 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T17:24:39,543 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T17:24:39,544 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T17:24:39,549 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T17:24:39,549 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T17:24:39,550 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T17:24:39,550 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T17:24:39,551 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,551 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T17:24:39,552 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,560 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T17:24:39,561 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T17:24:39,561 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T17:24:39,571 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T17:24:39,572 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T17:24:39,577 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160279573,5,FailOnTimeoutGroup] 2024-12-02T17:24:39,583 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160279578,5,FailOnTimeoutGroup] 2024-12-02T17:24:39,583 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,583 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T17:24:39,585 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,585 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:24:39,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:24:39,600 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T17:24:39,600 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0 2024-12-02T17:24:39,627 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:24:39,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:24:39,630 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56367, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T17:24:39,632 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:24:39,637 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:24:39,639 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37801 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 3b3375d1a925,38409,1733160278146 2024-12-02T17:24:39,642 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37801 {}] master.ServerManager(517): Registering regionserver=3b3375d1a925,38409,1733160278146 2024-12-02T17:24:39,647 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:24:39,647 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,648 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:39,649 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:24:39,653 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:24:39,653 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,655 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:39,656 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:24:39,659 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:24:39,659 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,660 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:39,661 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:24:39,664 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:24:39,664 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:39,666 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:39,666 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:24:39,667 DEBUG [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0 2024-12-02T17:24:39,667 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740 2024-12-02T17:24:39,667 DEBUG [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:38211 2024-12-02T17:24:39,668 DEBUG [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T17:24:39,668 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740 2024-12-02T17:24:39,673 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:24:39,674 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:24:39,674 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:24:39,675 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:24:39,676 DEBUG [RS:0;3b3375d1a925:38409 {}] zookeeper.ZKUtil(111): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/3b3375d1a925,38409,1733160278146 2024-12-02T17:24:39,677 WARN [RS:0;3b3375d1a925:38409 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:24:39,677 INFO [RS:0;3b3375d1a925:38409 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:24:39,677 DEBUG [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146 2024-12-02T17:24:39,680 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [3b3375d1a925,38409,1733160278146] 2024-12-02T17:24:39,681 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:24:39,698 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:24:39,699 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=833709, jitterRate=0.06011669337749481}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:24:39,704 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733160279632Initializing all the Stores at 1733160279636 (+4 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160279636Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160279636Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160279636Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160279636Cleaning up temporary data from old regions at 1733160279674 (+38 ms)Region opened successfully at 1733160279704 (+30 ms) 2024-12-02T17:24:39,704 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:24:39,704 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:24:39,705 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:24:39,705 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:24:39,705 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:24:39,709 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:24:39,709 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160279704Disabling compacts and flushes for region at 1733160279704Disabling writes for close at 1733160279705 (+1 ms)Writing region close event to WAL at 1733160279709 (+4 ms)Closed at 1733160279709 2024-12-02T17:24:39,715 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:24:39,715 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T17:24:39,721 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T17:24:39,724 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T17:24:39,738 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:24:39,744 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T17:24:39,750 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T17:24:39,757 INFO [RS:0;3b3375d1a925:38409 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:24:39,757 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,758 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T17:24:39,766 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T17:24:39,768 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,769 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,769 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,769 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,769 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,769 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,769 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:24:39,770 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,770 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,770 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,770 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,770 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,770 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:24:39,771 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:24:39,771 DEBUG [RS:0;3b3375d1a925:38409 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:24:39,784 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,785 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,785 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,785 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,786 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,786 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,38409,1733160278146-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:24:39,812 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T17:24:39,814 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,38409,1733160278146-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,815 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,815 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.Replication(171): 3b3375d1a925,38409,1733160278146 started 2024-12-02T17:24:39,837 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:39,837 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1482): Serving as 3b3375d1a925,38409,1733160278146, RpcServer on 3b3375d1a925/172.17.0.3:38409, sessionid=0x1009c051d0c0001 2024-12-02T17:24:39,838 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T17:24:39,838 DEBUG [RS:0;3b3375d1a925:38409 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 3b3375d1a925,38409,1733160278146 2024-12-02T17:24:39,839 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,38409,1733160278146' 2024-12-02T17:24:39,839 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T17:24:39,840 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T17:24:39,841 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T17:24:39,841 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T17:24:39,841 DEBUG [RS:0;3b3375d1a925:38409 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 3b3375d1a925,38409,1733160278146 2024-12-02T17:24:39,841 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,38409,1733160278146' 2024-12-02T17:24:39,841 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T17:24:39,842 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T17:24:39,843 DEBUG [RS:0;3b3375d1a925:38409 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T17:24:39,843 INFO [RS:0;3b3375d1a925:38409 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T17:24:39,843 INFO [RS:0;3b3375d1a925:38409 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T17:24:39,895 WARN [3b3375d1a925:37801 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-02T17:24:39,954 INFO [RS:0;3b3375d1a925:38409 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C38409%2C1733160278146, suffix=, logDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146, archiveDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs, maxLogs=32 2024-12-02T17:24:39,958 INFO [RS:0;3b3375d1a925:38409 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160279957 2024-12-02T17:24:39,970 INFO [RS:0;3b3375d1a925:38409 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160279957 2024-12-02T17:24:39,971 DEBUG [RS:0;3b3375d1a925:38409 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34997:34997),(127.0.0.1/127.0.0.1:33273:33273)] 2024-12-02T17:24:40,148 DEBUG [3b3375d1a925:37801 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T17:24:40,165 INFO [PEWorker-2 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=3b3375d1a925,38409,1733160278146 2024-12-02T17:24:40,174 INFO [PEWorker-2 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,38409,1733160278146, state=OPENING 2024-12-02T17:24:40,181 DEBUG [PEWorker-2 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T17:24:40,183 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:40,183 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:24:40,184 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:24:40,184 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:24:40,185 DEBUG [PEWorker-2 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:24:40,187 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,38409,1733160278146}] 2024-12-02T17:24:40,369 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T17:24:40,373 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:53091, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T17:24:40,388 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T17:24:40,389 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:24:40,395 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C38409%2C1733160278146.meta, suffix=.meta, logDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146, archiveDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs, maxLogs=32 2024-12-02T17:24:40,398 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.meta.1733160280398.meta 2024-12-02T17:24:40,413 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.meta.1733160280398.meta 2024-12-02T17:24:40,416 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:33273:33273),(127.0.0.1/127.0.0.1:34997:34997)] 2024-12-02T17:24:40,424 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:24:40,426 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T17:24:40,429 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T17:24:40,435 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T17:24:40,440 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T17:24:40,440 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:24:40,440 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T17:24:40,441 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T17:24:40,444 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:24:40,445 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:24:40,445 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:40,446 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:40,446 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:24:40,449 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:24:40,449 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:40,450 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:40,450 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:24:40,452 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:24:40,452 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:40,453 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:40,453 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:24:40,455 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:24:40,455 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:40,456 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:24:40,456 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:24:40,457 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740 2024-12-02T17:24:40,460 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740 2024-12-02T17:24:40,463 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:24:40,463 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:24:40,464 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:24:40,467 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:24:40,469 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=732528, jitterRate=-0.06854318082332611}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:24:40,469 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T17:24:40,471 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733160280441Writing region info on filesystem at 1733160280441Initializing all the Stores at 1733160280443 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160280443Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160280443Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160280443Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160280444 (+1 ms)Cleaning up temporary data from old regions at 1733160280463 (+19 ms)Running coprocessor post-open hooks at 1733160280469 (+6 ms)Region opened successfully at 1733160280471 (+2 ms) 2024-12-02T17:24:40,479 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733160280359 2024-12-02T17:24:40,496 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T17:24:40,497 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T17:24:40,499 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,38409,1733160278146 2024-12-02T17:24:40,501 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,38409,1733160278146, state=OPEN 2024-12-02T17:24:40,506 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:24:40,507 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:24:40,512 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:24:40,512 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:24:40,513 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=3b3375d1a925,38409,1733160278146 2024-12-02T17:24:40,519 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T17:24:40,519 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,38409,1733160278146 in 326 msec 2024-12-02T17:24:40,528 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T17:24:40,528 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 797 msec 2024-12-02T17:24:40,529 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:24:40,529 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T17:24:40,551 DEBUG [PEWorker-1 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:24:40,552 DEBUG [PEWorker-1 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,38409,1733160278146, seqNum=-1] 2024-12-02T17:24:40,580 DEBUG [PEWorker-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:24:40,582 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59047, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:24:40,607 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1710 sec 2024-12-02T17:24:40,607 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733160280607, completionTime=-1 2024-12-02T17:24:40,611 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T17:24:40,611 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T17:24:40,641 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=1 2024-12-02T17:24:40,641 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733160340641 2024-12-02T17:24:40,641 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733160400641 2024-12-02T17:24:40,641 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 29 msec 2024-12-02T17:24:40,644 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37801,1733160277462-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:40,644 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37801,1733160277462-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:40,644 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37801,1733160277462-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:40,646 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-3b3375d1a925:37801, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:40,646 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:40,647 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:40,655 DEBUG [master/3b3375d1a925:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T17:24:40,686 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.385sec 2024-12-02T17:24:40,687 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T17:24:40,689 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T17:24:40,690 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T17:24:40,690 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T17:24:40,690 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T17:24:40,691 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37801,1733160277462-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:24:40,692 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37801,1733160277462-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T17:24:40,705 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T17:24:40,707 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T17:24:40,707 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37801,1733160277462-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:24:40,767 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4731d90b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:24:40,770 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-02T17:24:40,770 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-02T17:24:40,773 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 3b3375d1a925,37801,-1 for getting cluster id 2024-12-02T17:24:40,777 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T17:24:40,787 DEBUG [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'ee9ea7b4-ea45-493c-bef1-e3bf24bf530b' 2024-12-02T17:24:40,790 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T17:24:40,790 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "ee9ea7b4-ea45-493c-bef1-e3bf24bf530b" 2024-12-02T17:24:40,793 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@34d065df, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:24:40,793 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [3b3375d1a925,37801,-1] 2024-12-02T17:24:40,795 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T17:24:40,797 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:24:40,799 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45370, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T17:24:40,804 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2eb3700d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:24:40,805 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:24:40,813 DEBUG [RPCClient-NioEventLoopGroup-4-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,38409,1733160278146, seqNum=-1] 2024-12-02T17:24:40,814 DEBUG [RPCClient-NioEventLoopGroup-4-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:24:40,817 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42462, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:24:40,839 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=3b3375d1a925,37801,1733160277462 2024-12-02T17:24:40,839 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:24:40,847 INFO [Time-limited test {}] master.MasterRpcServices(567): Client=null/null set balanceSwitch=false 2024-12-02T17:24:40,851 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-02T17:24:40,855 DEBUG [RPCClient-NioEventLoopGroup-4-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 3b3375d1a925,37801,1733160277462 2024-12-02T17:24:40,859 DEBUG [RPCClient-NioEventLoopGroup-4-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@4cfe2c38 2024-12-02T17:24:40,860 DEBUG [RPCClient-NioEventLoopGroup-4-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T17:24:40,864 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45384, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T17:24:40,867 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37801 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T17:24:40,867 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37801 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T17:24:40,871 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37801 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestLogRolling-testSlowSyncLogRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:24:40,880 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37801 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling 2024-12-02T17:24:40,883 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T17:24:40,885 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37801 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testSlowSyncLogRolling" procId is: 4 2024-12-02T17:24:40,886 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:40,889 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T17:24:40,891 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37801 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:24:40,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741835_1011 (size=389) 2024-12-02T17:24:40,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741835_1011 (size=389) 2024-12-02T17:24:40,954 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 247883c968739afd77f6940a11f2ce1d, NAME => 'TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testSlowSyncLogRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0 2024-12-02T17:24:40,972 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741836_1012 (size=72) 2024-12-02T17:24:40,972 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741836_1012 (size=72) 2024-12-02T17:24:41,374 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(898): Instantiated TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:24:41,374 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1722): Closing 247883c968739afd77f6940a11f2ce1d, disabling compactions & flushes 2024-12-02T17:24:41,374 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1755): Closing region TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:24:41,374 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:24:41,374 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. after waiting 0 ms 2024-12-02T17:24:41,374 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:24:41,374 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1973): Closed TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:24:41,374 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1676): Region close journal for 247883c968739afd77f6940a11f2ce1d: Waiting for close lock at 1733160281374Disabling compacts and flushes for region at 1733160281374Disabling writes for close at 1733160281374Writing region close event to WAL at 1733160281374Closed at 1733160281374 2024-12-02T17:24:41,377 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T17:24:41,383 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.","families":{"info":[{"qualifier":"regioninfo","vlen":71,"tag":[],"timestamp":"1733160281377"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733160281377"}]},"ts":"1733160281377"} 2024-12-02T17:24:41,388 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-02T17:24:41,390 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T17:24:41,392 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testSlowSyncLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160281390"}]},"ts":"1733160281390"} 2024-12-02T17:24:41,397 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testSlowSyncLogRolling, state=ENABLING in hbase:meta 2024-12-02T17:24:41,399 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=247883c968739afd77f6940a11f2ce1d, ASSIGN}] 2024-12-02T17:24:41,401 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=247883c968739afd77f6940a11f2ce1d, ASSIGN 2024-12-02T17:24:41,403 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=247883c968739afd77f6940a11f2ce1d, ASSIGN; state=OFFLINE, location=3b3375d1a925,38409,1733160278146; forceNewPlan=false, retain=false 2024-12-02T17:24:41,554 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=247883c968739afd77f6940a11f2ce1d, regionState=OPENING, regionLocation=3b3375d1a925,38409,1733160278146 2024-12-02T17:24:41,558 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=247883c968739afd77f6940a11f2ce1d, ASSIGN because future has completed 2024-12-02T17:24:41,559 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 247883c968739afd77f6940a11f2ce1d, server=3b3375d1a925,38409,1733160278146}] 2024-12-02T17:24:41,719 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:24:41,719 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 247883c968739afd77f6940a11f2ce1d, NAME => 'TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:24:41,720 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testSlowSyncLogRolling 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,720 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:24:41,720 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,720 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,723 INFO [StoreOpener-247883c968739afd77f6940a11f2ce1d-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,725 INFO [StoreOpener-247883c968739afd77f6940a11f2ce1d-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 247883c968739afd77f6940a11f2ce1d columnFamilyName info 2024-12-02T17:24:41,725 DEBUG [StoreOpener-247883c968739afd77f6940a11f2ce1d-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:24:41,727 INFO [StoreOpener-247883c968739afd77f6940a11f2ce1d-1 {}] regionserver.HStore(327): Store=247883c968739afd77f6940a11f2ce1d/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:24:41,727 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,728 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,729 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,729 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,729 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,732 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,736 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:24:41,737 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 247883c968739afd77f6940a11f2ce1d; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=857665, jitterRate=0.09057851135730743}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T17:24:41,737 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:24:41,738 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 247883c968739afd77f6940a11f2ce1d: Running coprocessor pre-open hook at 1733160281720Writing region info on filesystem at 1733160281720Initializing all the Stores at 1733160281722 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160281722Cleaning up temporary data from old regions at 1733160281730 (+8 ms)Running coprocessor post-open hooks at 1733160281737 (+7 ms)Region opened successfully at 1733160281738 (+1 ms) 2024-12-02T17:24:41,740 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d., pid=6, masterSystemTime=1733160281712 2024-12-02T17:24:41,744 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:24:41,745 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:24:41,746 INFO [PEWorker-2 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=247883c968739afd77f6940a11f2ce1d, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,38409,1733160278146 2024-12-02T17:24:41,750 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 247883c968739afd77f6940a11f2ce1d, server=3b3375d1a925,38409,1733160278146 because future has completed 2024-12-02T17:24:41,755 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T17:24:41,756 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 247883c968739afd77f6940a11f2ce1d, server=3b3375d1a925,38409,1733160278146 in 193 msec 2024-12-02T17:24:41,760 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T17:24:41,760 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=247883c968739afd77f6940a11f2ce1d, ASSIGN in 357 msec 2024-12-02T17:24:41,761 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T17:24:41,761 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testSlowSyncLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160281761"}]},"ts":"1733160281761"} 2024-12-02T17:24:41,765 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testSlowSyncLogRolling, state=ENABLED in hbase:meta 2024-12-02T17:24:41,767 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T17:24:41,770 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling in 893 msec 2024-12-02T17:24:45,944 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-hbase.properties,hadoop-metrics2.properties 2024-12-02T17:24:46,029 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T17:24:46,031 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testSlowSyncLogRolling' 2024-12-02T17:24:47,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:24:47,868 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-02T17:24:47,869 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling 2024-12-02T17:24:47,869 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling Metrics about Tables on a single HBase RegionServer 2024-12-02T17:24:47,870 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:24:47,870 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-02T17:24:47,870 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-02T17:24:47,871 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-02T17:24:50,929 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37801 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:24:50,931 INFO [RPCClient-NioEventLoopGroup-4-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestLogRolling-testSlowSyncLogRolling completed 2024-12-02T17:24:50,934 DEBUG [Time-limited test {}] hbase.ClientMetaTableAccessor(255): Scanning META starting at row=TestLogRolling-testSlowSyncLogRolling,, stopping at row=TestLogRolling-testSlowSyncLogRolling ,, for max=2147483647 with caching=100 2024-12-02T17:24:50,940 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2234): Found 1 regions for table TestLogRolling-testSlowSyncLogRolling 2024-12-02T17:24:50,941 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2240): firstRegionName=TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:24:50,942 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160290941 2024-12-02T17:24:50,950 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:24:50,950 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:24:50,950 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:24:50,950 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:24:50,950 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:24:50,951 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160279957 with entries=1, filesize=443 B; new WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160290941 2024-12-02T17:24:50,952 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:33273:33273),(127.0.0.1/127.0.0.1:34997:34997)] 2024-12-02T17:24:50,952 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160279957 is not closed yet, will try archiving it next time 2024-12-02T17:24:50,954 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741833_1009 (size=451) 2024-12-02T17:24:50,954 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741833_1009 (size=451) 2024-12-02T17:24:50,958 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160279957 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs/3b3375d1a925%2C38409%2C1733160278146.1733160279957 2024-12-02T17:24:50,960 DEBUG [RPCClient-NioEventLoopGroup-4-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestLogRolling-testSlowSyncLogRolling', row='row0001', locateType=CURRENT is [region=TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d., hostname=3b3375d1a925,38409,1733160278146, seqNum=2] 2024-12-02T17:25:02,996 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38409 {}] regionserver.HRegion(8855): Flush requested on 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:25:03,001 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 247883c968739afd77f6940a11f2ce1d 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:25:03,072 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/c9f7808de56548aa8ec61516929af3d2 is 1080, key is row0001/info:/1733160290963/Put/seqid=0 2024-12-02T17:25:03,085 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741838_1014 (size=12509) 2024-12-02T17:25:03,086 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741838_1014 (size=12509) 2024-12-02T17:25:03,087 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/c9f7808de56548aa8ec61516929af3d2 2024-12-02T17:25:03,142 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/c9f7808de56548aa8ec61516929af3d2 as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c9f7808de56548aa8ec61516929af3d2 2024-12-02T17:25:03,153 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c9f7808de56548aa8ec61516929af3d2, entries=7, sequenceid=11, filesize=12.2 K 2024-12-02T17:25:03,160 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 247883c968739afd77f6940a11f2ce1d in 160ms, sequenceid=11, compaction requested=false 2024-12-02T17:25:03,161 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 247883c968739afd77f6940a11f2ce1d: 2024-12-02T17:25:06,512 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:25:11,010 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160311009 2024-12-02T17:25:11,218 INFO [Time-limited test {}] wal.AbstractFSWAL(1368): Slow sync cost: 205 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:11,218 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:11,219 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:11,219 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:11,219 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:11,219 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:11,219 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160290941 with entries=12, filesize=12.10 KB; new WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160311009 2024-12-02T17:25:11,221 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34997:34997),(127.0.0.1/127.0.0.1:33273:33273)] 2024-12-02T17:25:11,221 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160290941 is not closed yet, will try archiving it next time 2024-12-02T17:25:11,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741837_1013 (size=12399) 2024-12-02T17:25:11,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741837_1013 (size=12399) 2024-12-02T17:25:11,424 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:13,627 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:15,832 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:18,035 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:18,036 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38409 {}] regionserver.HRegion(8855): Flush requested on 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:25:18,036 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 247883c968739afd77f6940a11f2ce1d 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:25:18,238 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:18,243 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/4f53ea2437334862b17c93953e3cdb85 is 1080, key is row0008/info:/1733160304998/Put/seqid=0 2024-12-02T17:25:18,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741840_1016 (size=12509) 2024-12-02T17:25:18,254 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741840_1016 (size=12509) 2024-12-02T17:25:18,255 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=21 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/4f53ea2437334862b17c93953e3cdb85 2024-12-02T17:25:18,264 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/4f53ea2437334862b17c93953e3cdb85 as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/4f53ea2437334862b17c93953e3cdb85 2024-12-02T17:25:18,272 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/4f53ea2437334862b17c93953e3cdb85, entries=7, sequenceid=21, filesize=12.2 K 2024-12-02T17:25:18,474 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:18,474 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 247883c968739afd77f6940a11f2ce1d in 438ms, sequenceid=21, compaction requested=false 2024-12-02T17:25:18,475 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 247883c968739afd77f6940a11f2ce1d: 2024-12-02T17:25:18,475 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=24.4 K, sizeToCheck=16.0 K 2024-12-02T17:25:18,475 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:25:18,476 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c9f7808de56548aa8ec61516929af3d2 because midkey is the same as first or last row 2024-12-02T17:25:20,240 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:21,536 INFO [master/3b3375d1a925:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T17:25:21,536 INFO [master/3b3375d1a925:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T17:25:22,444 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:22,446 WARN [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(2201): Requesting log roll because we exceeded slow sync threshold; count=8, threshold=5, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:22,447 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C38409%2C1733160278146:(num 1733160311009) roll requested 2024-12-02T17:25:22,448 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160322447 2024-12-02T17:25:22,655 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1368): Slow sync cost: 205 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:22,656 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:22,656 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:22,656 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:22,656 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:22,656 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:22,656 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160311009 with entries=8, filesize=7.55 KB; new WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160322447 2024-12-02T17:25:22,658 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:33273:33273),(127.0.0.1/127.0.0.1:34997:34997)] 2024-12-02T17:25:22,658 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160311009 is not closed yet, will try archiving it next time 2024-12-02T17:25:22,658 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160290941 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs/3b3375d1a925%2C38409%2C1733160278146.1733160290941 2024-12-02T17:25:22,659 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741839_1015 (size=7739) 2024-12-02T17:25:22,659 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741839_1015 (size=7739) 2024-12-02T17:25:24,647 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:26,720 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(329): Region 247883c968739afd77f6940a11f2ce1d, had cached 0 bytes from a total of 25018 2024-12-02T17:25:26,851 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:29,055 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:31,259 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:33,261 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [10,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T17:25:33,261 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160333261 2024-12-02T17:25:36,512 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:25:38,270 INFO [Time-limited test {}] wal.AbstractFSWAL(1368): Slow sync cost: 5006 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:38,272 WARN [Time-limited test {}] wal.AbstractFSWAL(1374): Requesting log roll because we exceeded slow sync threshold; time=5006 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:38,272 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C38409%2C1733160278146:(num 1733160333261) roll requested 2024-12-02T17:25:38,273 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:38,273 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:38,273 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:38,273 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:38,273 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:38,273 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160322447 with entries=4, filesize=4.63 KB; new WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160333261 2024-12-02T17:25:38,276 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741841_1017 (size=4753) 2024-12-02T17:25:38,276 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741841_1017 (size=4753) 2024-12-02T17:25:38,285 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34997:34997),(127.0.0.1/127.0.0.1:33273:33273)] 2024-12-02T17:25:38,285 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160322447 is not closed yet, will try archiving it next time 2024-12-02T17:25:38,285 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160338285 2024-12-02T17:25:43,289 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 5001 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:43,289 WARN [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1374): Requesting log roll because we exceeded slow sync threshold; time=5001 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:43,289 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38409 {}] regionserver.HRegion(8855): Flush requested on 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:25:43,289 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 247883c968739afd77f6940a11f2ce1d 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:25:43,295 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1368): Slow sync cost: 5006 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:43,295 WARN [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1374): Requesting log roll because we exceeded slow sync threshold; time=5006 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:45,290 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [10,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T17:25:48,292 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 5001 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:48,292 WARN [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1374): Requesting log roll because we exceeded slow sync threshold; time=5001 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK], DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK]] 2024-12-02T17:25:48,292 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:48,292 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:48,292 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:48,292 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:48,293 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:48,293 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160333261 with entries=2, filesize=1.52 KB; new WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160338285 2024-12-02T17:25:48,294 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:33273:33273),(127.0.0.1/127.0.0.1:34997:34997)] 2024-12-02T17:25:48,294 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160333261 is not closed yet, will try archiving it next time 2024-12-02T17:25:48,294 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C38409%2C1733160278146:(num 1733160338285) roll requested 2024-12-02T17:25:48,294 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160348294 2024-12-02T17:25:48,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741842_1018 (size=1569) 2024-12-02T17:25:48,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741842_1018 (size=1569) 2024-12-02T17:25:48,299 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/3ad4175ca8384a68a22eec485bf51c62 is 1080, key is row0015/info:/1733160320038/Put/seqid=0 2024-12-02T17:25:48,306 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741844_1020 (size=12509) 2024-12-02T17:25:48,307 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741844_1020 (size=12509) 2024-12-02T17:25:48,307 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=31 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/3ad4175ca8384a68a22eec485bf51c62 2024-12-02T17:25:48,318 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/3ad4175ca8384a68a22eec485bf51c62 as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/3ad4175ca8384a68a22eec485bf51c62 2024-12-02T17:25:48,327 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/3ad4175ca8384a68a22eec485bf51c62, entries=7, sequenceid=31, filesize=12.2 K 2024-12-02T17:25:53,302 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1368): Slow sync cost: 5005 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:53,302 WARN [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1374): Requesting log roll because we exceeded slow sync threshold; time=5005 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:53,329 INFO [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1368): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:53,329 WARN [FSHLog-0-hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0-prefix:3b3375d1a925,38409,1733160278146 {}] wal.AbstractFSWAL(1374): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44647,DS-41c49996-612d-45af-b6c8-724ff3d6f8f0,DISK], DatanodeInfoWithStorage[127.0.0.1:33621,DS-ba5b9e21-f98a-46e3-8092-1be1efcd0579,DISK]] 2024-12-02T17:25:53,329 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 247883c968739afd77f6940a11f2ce1d in 10040ms, sequenceid=31, compaction requested=true 2024-12-02T17:25:53,329 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,329 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 247883c968739afd77f6940a11f2ce1d: 2024-12-02T17:25:53,329 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,329 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,329 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=36.6 K, sizeToCheck=16.0 K 2024-12-02T17:25:53,329 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:25:53,329 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,329 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c9f7808de56548aa8ec61516929af3d2 because midkey is the same as first or last row 2024-12-02T17:25:53,330 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,330 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160338285 with entries=1, filesize=430 B; new WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160348294 2024-12-02T17:25:53,330 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34997:34997),(127.0.0.1/127.0.0.1:33273:33273)] 2024-12-02T17:25:53,331 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160338285 is not closed yet, will try archiving it next time 2024-12-02T17:25:53,331 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160311009 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs/3b3375d1a925%2C38409%2C1733160278146.1733160311009 2024-12-02T17:25:53,331 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C38409%2C1733160278146:(num 1733160348294) roll requested 2024-12-02T17:25:53,331 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160353331 2024-12-02T17:25:53,331 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 247883c968739afd77f6940a11f2ce1d:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:25:53,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741843_1019 (size=438) 2024-12-02T17:25:53,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741843_1019 (size=438) 2024-12-02T17:25:53,334 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:25:53,334 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:25:53,335 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160322447 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs/3b3375d1a925%2C38409%2C1733160278146.1733160322447 2024-12-02T17:25:53,337 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160333261 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs/3b3375d1a925%2C38409%2C1733160278146.1733160333261 2024-12-02T17:25:53,337 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 37527 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:25:53,339 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.HStore(1541): 247883c968739afd77f6940a11f2ce1d/info is initiating minor compaction (all files) 2024-12-02T17:25:53,339 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160338285 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs/3b3375d1a925%2C38409%2C1733160278146.1733160338285 2024-12-02T17:25:53,339 INFO [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of 247883c968739afd77f6940a11f2ce1d/info in TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:25:53,340 INFO [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c9f7808de56548aa8ec61516929af3d2, hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/4f53ea2437334862b17c93953e3cdb85, hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/3ad4175ca8384a68a22eec485bf51c62] into tmpdir=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp, totalSize=36.6 K 2024-12-02T17:25:53,341 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] compactions.Compactor(225): Compacting c9f7808de56548aa8ec61516929af3d2, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=11, earliestPutTs=1733160290963 2024-12-02T17:25:53,342 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] compactions.Compactor(225): Compacting 4f53ea2437334862b17c93953e3cdb85, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=21, earliestPutTs=1733160304998 2024-12-02T17:25:53,342 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] compactions.Compactor(225): Compacting 3ad4175ca8384a68a22eec485bf51c62, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=31, earliestPutTs=1733160320038 2024-12-02T17:25:53,346 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,347 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,347 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,347 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,347 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,347 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160348294 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160353331 2024-12-02T17:25:53,349 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34997:34997),(127.0.0.1/127.0.0.1:33273:33273)] 2024-12-02T17:25:53,349 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160348294 is not closed yet, will try archiving it next time 2024-12-02T17:25:53,349 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741845_1021 (size=93) 2024-12-02T17:25:53,349 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C38409%2C1733160278146.1733160353349 2024-12-02T17:25:53,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741845_1021 (size=93) 2024-12-02T17:25:53,351 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160348294 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs/3b3375d1a925%2C38409%2C1733160278146.1733160348294 2024-12-02T17:25:53,362 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,362 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,362 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,362 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,363 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:25:53,363 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160353331 with entries=1, filesize=1.22 KB; new WAL /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/WALs/3b3375d1a925,38409,1733160278146/3b3375d1a925%2C38409%2C1733160278146.1733160353349 2024-12-02T17:25:53,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741846_1022 (size=1258) 2024-12-02T17:25:53,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741846_1022 (size=1258) 2024-12-02T17:25:53,369 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34997:34997),(127.0.0.1/127.0.0.1:33273:33273)] 2024-12-02T17:25:53,378 INFO [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 247883c968739afd77f6940a11f2ce1d#info#compaction#3 average throughput is 10.77 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:25:53,379 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/c68a3b940f6c49dc9cf989c64cc5c000 is 1080, key is row0001/info:/1733160290963/Put/seqid=0 2024-12-02T17:25:53,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741848_1024 (size=27710) 2024-12-02T17:25:53,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741848_1024 (size=27710) 2024-12-02T17:25:53,399 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/c68a3b940f6c49dc9cf989c64cc5c000 as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c68a3b940f6c49dc9cf989c64cc5c000 2024-12-02T17:25:53,417 INFO [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in 247883c968739afd77f6940a11f2ce1d/info of 247883c968739afd77f6940a11f2ce1d into c68a3b940f6c49dc9cf989c64cc5c000(size=27.1 K), total size for store is 27.1 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:25:53,417 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for 247883c968739afd77f6940a11f2ce1d: 2024-12-02T17:25:53,420 INFO [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d., storeName=247883c968739afd77f6940a11f2ce1d/info, priority=13, startTime=1733160353331; duration=0sec 2024-12-02T17:25:53,420 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=27.1 K, sizeToCheck=16.0 K 2024-12-02T17:25:53,420 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:25:53,420 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c68a3b940f6c49dc9cf989c64cc5c000 because midkey is the same as first or last row 2024-12-02T17:25:53,420 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=27.1 K, sizeToCheck=16.0 K 2024-12-02T17:25:53,420 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:25:53,421 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c68a3b940f6c49dc9cf989c64cc5c000 because midkey is the same as first or last row 2024-12-02T17:25:53,421 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=27.1 K, sizeToCheck=16.0 K 2024-12-02T17:25:53,421 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:25:53,421 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c68a3b940f6c49dc9cf989c64cc5c000 because midkey is the same as first or last row 2024-12-02T17:25:53,421 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:25:53,421 DEBUG [RS:0;3b3375d1a925:38409-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 247883c968739afd77f6940a11f2ce1d:info 2024-12-02T17:26:05,378 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38409 {}] regionserver.HRegion(8855): Flush requested on 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:26:05,378 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 247883c968739afd77f6940a11f2ce1d 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:26:05,385 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/97199e17b78d4e01ac8fa0a4b0b2bf23 is 1080, key is row0022/info:/1733160353351/Put/seqid=0 2024-12-02T17:26:05,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741849_1025 (size=12509) 2024-12-02T17:26:05,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741849_1025 (size=12509) 2024-12-02T17:26:05,393 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=42 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/97199e17b78d4e01ac8fa0a4b0b2bf23 2024-12-02T17:26:05,402 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/97199e17b78d4e01ac8fa0a4b0b2bf23 as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/97199e17b78d4e01ac8fa0a4b0b2bf23 2024-12-02T17:26:05,409 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/97199e17b78d4e01ac8fa0a4b0b2bf23, entries=7, sequenceid=42, filesize=12.2 K 2024-12-02T17:26:05,410 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 247883c968739afd77f6940a11f2ce1d in 32ms, sequenceid=42, compaction requested=false 2024-12-02T17:26:05,411 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 247883c968739afd77f6940a11f2ce1d: 2024-12-02T17:26:05,411 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=39.3 K, sizeToCheck=16.0 K 2024-12-02T17:26:05,411 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:05,411 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c68a3b940f6c49dc9cf989c64cc5c000 because midkey is the same as first or last row 2024-12-02T17:26:06,512 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:26:11,721 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(329): Region 247883c968739afd77f6940a11f2ce1d, had cached 0 bytes from a total of 40219 2024-12-02T17:26:13,389 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T17:26:13,390 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:26:13,390 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:13,395 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:13,395 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:13,395 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T17:26:13,396 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T17:26:13,396 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=933253213, stopped=false 2024-12-02T17:26:13,396 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=3b3375d1a925,37801,1733160277462 2024-12-02T17:26:13,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:13,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:13,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:13,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:13,399 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:26:13,399 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:26:13,399 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:13,399 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:13,400 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:13,400 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '3b3375d1a925,38409,1733160278146' ***** 2024-12-02T17:26:13,400 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:13,400 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T17:26:13,400 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T17:26:13,401 INFO [RS:0;3b3375d1a925:38409 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T17:26:13,401 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T17:26:13,401 INFO [RS:0;3b3375d1a925:38409 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T17:26:13,401 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(3091): Received CLOSE for 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:26:13,402 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(959): stopping server 3b3375d1a925,38409,1733160278146 2024-12-02T17:26:13,402 INFO [RS:0;3b3375d1a925:38409 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:26:13,402 INFO [RS:0;3b3375d1a925:38409 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;3b3375d1a925:38409. 2024-12-02T17:26:13,402 DEBUG [RS:0;3b3375d1a925:38409 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:13,402 DEBUG [RS:0;3b3375d1a925:38409 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:13,402 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 247883c968739afd77f6940a11f2ce1d, disabling compactions & flushes 2024-12-02T17:26:13,402 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T17:26:13,402 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:26:13,402 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T17:26:13,402 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:26:13,402 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T17:26:13,402 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. after waiting 0 ms 2024-12-02T17:26:13,402 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:26:13,402 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T17:26:13,403 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2902): Flushing 247883c968739afd77f6940a11f2ce1d 1/1 column families, dataSize=3.15 KB heapSize=3.63 KB 2024-12-02T17:26:13,403 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-02T17:26:13,403 DEBUG [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1325): Online Regions={247883c968739afd77f6940a11f2ce1d=TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d., 1588230740=hbase:meta,,1.1588230740} 2024-12-02T17:26:13,403 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:26:13,403 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:26:13,403 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:26:13,403 DEBUG [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, 247883c968739afd77f6940a11f2ce1d 2024-12-02T17:26:13,403 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:26:13,403 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:26:13,404 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.65 KB heapSize=3.67 KB 2024-12-02T17:26:13,409 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/8ec14d51905640b2901bd35e257e0bbf is 1080, key is row0029/info:/1733160367380/Put/seqid=0 2024-12-02T17:26:13,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741850_1026 (size=8193) 2024-12-02T17:26:13,418 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741850_1026 (size=8193) 2024-12-02T17:26:13,419 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=3.15 KB at sequenceid=48 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/8ec14d51905640b2901bd35e257e0bbf 2024-12-02T17:26:13,428 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/.tmp/info/8ec14d51905640b2901bd35e257e0bbf as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/8ec14d51905640b2901bd35e257e0bbf 2024-12-02T17:26:13,429 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/info/19180fdef56b467aafd9e800ed75d96d is 195, key is TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d./info:regioninfo/1733160281745/Put/seqid=0 2024-12-02T17:26:13,436 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/8ec14d51905640b2901bd35e257e0bbf, entries=3, sequenceid=48, filesize=8.0 K 2024-12-02T17:26:13,437 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3140): Finished flush of dataSize ~3.15 KB/3228, heapSize ~3.61 KB/3696, currentSize=0 B/0 for 247883c968739afd77f6940a11f2ce1d in 35ms, sequenceid=48, compaction requested=true 2024-12-02T17:26:13,438 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741851_1027 (size=7016) 2024-12-02T17:26:13,438 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741851_1027 (size=7016) 2024-12-02T17:26:13,439 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.45 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/info/19180fdef56b467aafd9e800ed75d96d 2024-12-02T17:26:13,445 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.-1 {}] regionserver.HStore(2317): Moving the files [hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c9f7808de56548aa8ec61516929af3d2, hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/4f53ea2437334862b17c93953e3cdb85, hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/3ad4175ca8384a68a22eec485bf51c62] to archive 2024-12-02T17:26:13,448 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T17:26:13,451 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c9f7808de56548aa8ec61516929af3d2 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/archive/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/c9f7808de56548aa8ec61516929af3d2 2024-12-02T17:26:13,454 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/4f53ea2437334862b17c93953e3cdb85 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/archive/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/4f53ea2437334862b17c93953e3cdb85 2024-12-02T17:26:13,456 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/3ad4175ca8384a68a22eec485bf51c62 to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/archive/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/info/3ad4175ca8384a68a22eec485bf51c62 2024-12-02T17:26:13,468 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/ns/184a9fce89be44e9b7b475e43e889672 is 43, key is default/ns:d/1733160280586/Put/seqid=0 2024-12-02T17:26:13,468 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.-1 {}] regionserver.HRegionServer(3532): Failed to report file archival(s) to Master. This will be retried. org.apache.hadoop.hbase.ipc.StoppedRpcClientException: Call to address=3b3375d1a925:37801 failed on local exception: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.hbase.ipc.IPCUtil.wrapException(IPCUtil.java:237) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:395) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:430) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:425) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:117) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.setException(Call.java:132) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:451) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:336) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:606) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$BlockingStub.reportFileArchival(RegionServerStatusProtos.java:17350) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegionServer.reportFileArchivalForQuotas(HRegionServer.java:3516) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.reportArchivedFilesForQuota(HStore.java:2412) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.removeCompactedfiles(HStore.java:2347) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.closeWithoutLock(HStore.java:738) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.close(HStore.java:804) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion$2.call(HRegion.java:1912) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion$2.call(HRegion.java:1909) ~[classes/:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) ~[?:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at org.apache.hadoop.hbase.ipc.AbstractRpcClient.getConnection(AbstractRpcClient.java:366) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:448) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] ... 16 more 2024-12-02T17:26:13,473 WARN [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d.-1 {}] regionserver.HStore(2414): Failed to report archival of files: [c9f7808de56548aa8ec61516929af3d2=12509, 4f53ea2437334862b17c93953e3cdb85=12509, 3ad4175ca8384a68a22eec485bf51c62=12509] 2024-12-02T17:26:13,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741852_1028 (size=5153) 2024-12-02T17:26:13,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741852_1028 (size=5153) 2024-12-02T17:26:13,477 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/ns/184a9fce89be44e9b7b475e43e889672 2024-12-02T17:26:13,480 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/default/TestLogRolling-testSlowSyncLogRolling/247883c968739afd77f6940a11f2ce1d/recovered.edits/51.seqid, newMaxSeqId=51, maxSeqId=1 2024-12-02T17:26:13,482 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:26:13,483 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 247883c968739afd77f6940a11f2ce1d: Waiting for close lock at 1733160373402Running coprocessor pre-close hooks at 1733160373402Disabling compacts and flushes for region at 1733160373402Disabling writes for close at 1733160373402Obtaining lock to block concurrent updates at 1733160373403 (+1 ms)Preparing flush snapshotting stores in 247883c968739afd77f6940a11f2ce1d at 1733160373403Finished memstore snapshotting TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d., syncing WAL and waiting on mvcc, flushsize=dataSize=3228, getHeapSize=3696, getOffHeapSize=0, getCellsCount=3 at 1733160373403Flushing stores of TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. at 1733160373404 (+1 ms)Flushing 247883c968739afd77f6940a11f2ce1d/info: creating writer at 1733160373404Flushing 247883c968739afd77f6940a11f2ce1d/info: appending metadata at 1733160373408 (+4 ms)Flushing 247883c968739afd77f6940a11f2ce1d/info: closing flushed file at 1733160373408Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4ca6b474: reopening flushed file at 1733160373427 (+19 ms)Finished flush of dataSize ~3.15 KB/3228, heapSize ~3.61 KB/3696, currentSize=0 B/0 for 247883c968739afd77f6940a11f2ce1d in 35ms, sequenceid=48, compaction requested=true at 1733160373437 (+10 ms)Writing region close event to WAL at 1733160373475 (+38 ms)Running coprocessor post-close hooks at 1733160373481 (+6 ms)Closed at 1733160373482 (+1 ms) 2024-12-02T17:26:13,483 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testSlowSyncLogRolling,,1733160280866.247883c968739afd77f6940a11f2ce1d. 2024-12-02T17:26:13,503 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/table/b01ba7cbcc6048cb8cbdcf47ee4fa46d is 73, key is TestLogRolling-testSlowSyncLogRolling/table:state/1733160281761/Put/seqid=0 2024-12-02T17:26:13,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741853_1029 (size=5396) 2024-12-02T17:26:13,510 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741853_1029 (size=5396) 2024-12-02T17:26:13,511 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=138 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/table/b01ba7cbcc6048cb8cbdcf47ee4fa46d 2024-12-02T17:26:13,521 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/info/19180fdef56b467aafd9e800ed75d96d as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/info/19180fdef56b467aafd9e800ed75d96d 2024-12-02T17:26:13,529 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/info/19180fdef56b467aafd9e800ed75d96d, entries=10, sequenceid=11, filesize=6.9 K 2024-12-02T17:26:13,530 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/ns/184a9fce89be44e9b7b475e43e889672 as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/ns/184a9fce89be44e9b7b475e43e889672 2024-12-02T17:26:13,539 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/ns/184a9fce89be44e9b7b475e43e889672, entries=2, sequenceid=11, filesize=5.0 K 2024-12-02T17:26:13,540 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/.tmp/table/b01ba7cbcc6048cb8cbdcf47ee4fa46d as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/table/b01ba7cbcc6048cb8cbdcf47ee4fa46d 2024-12-02T17:26:13,548 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/table/b01ba7cbcc6048cb8cbdcf47ee4fa46d, entries=2, sequenceid=11, filesize=5.3 K 2024-12-02T17:26:13,550 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.65 KB/1692, heapSize ~3.38 KB/3456, currentSize=0 B/0 for 1588230740 in 146ms, sequenceid=11, compaction requested=false 2024-12-02T17:26:13,556 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-02T17:26:13,557 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:26:13,557 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:26:13,557 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160373403Running coprocessor pre-close hooks at 1733160373403Disabling compacts and flushes for region at 1733160373403Disabling writes for close at 1733160373403Obtaining lock to block concurrent updates at 1733160373404 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1733160373404Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1692, getHeapSize=3696, getOffHeapSize=0, getCellsCount=14 at 1733160373404Flushing stores of hbase:meta,,1.1588230740 at 1733160373405 (+1 ms)Flushing 1588230740/info: creating writer at 1733160373405Flushing 1588230740/info: appending metadata at 1733160373429 (+24 ms)Flushing 1588230740/info: closing flushed file at 1733160373429Flushing 1588230740/ns: creating writer at 1733160373446 (+17 ms)Flushing 1588230740/ns: appending metadata at 1733160373468 (+22 ms)Flushing 1588230740/ns: closing flushed file at 1733160373468Flushing 1588230740/table: creating writer at 1733160373485 (+17 ms)Flushing 1588230740/table: appending metadata at 1733160373502 (+17 ms)Flushing 1588230740/table: closing flushed file at 1733160373502Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@12ca7b34: reopening flushed file at 1733160373519 (+17 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1a018d60: reopening flushed file at 1733160373529 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3fea8b0f: reopening flushed file at 1733160373539 (+10 ms)Finished flush of dataSize ~1.65 KB/1692, heapSize ~3.38 KB/3456, currentSize=0 B/0 for 1588230740 in 146ms, sequenceid=11, compaction requested=false at 1733160373550 (+11 ms)Writing region close event to WAL at 1733160373551 (+1 ms)Running coprocessor post-close hooks at 1733160373556 (+5 ms)Closed at 1733160373557 (+1 ms) 2024-12-02T17:26:13,557 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T17:26:13,604 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(976): stopping server 3b3375d1a925,38409,1733160278146; all regions closed. 2024-12-02T17:26:13,605 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,605 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,606 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,606 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,606 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,609 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741834_1010 (size=3066) 2024-12-02T17:26:13,609 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741834_1010 (size=3066) 2024-12-02T17:26:13,613 DEBUG [RS:0;3b3375d1a925:38409 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs 2024-12-02T17:26:13,613 INFO [RS:0;3b3375d1a925:38409 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C38409%2C1733160278146.meta:.meta(num 1733160280398) 2024-12-02T17:26:13,614 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,614 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,614 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,614 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,615 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,617 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741847_1023 (size=12695) 2024-12-02T17:26:13,620 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741847_1023 (size=12695) 2024-12-02T17:26:13,623 DEBUG [RS:0;3b3375d1a925:38409 {}] wal.AbstractFSWAL(1256): Moved 2 WAL file(s) to /user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/oldWALs 2024-12-02T17:26:13,624 INFO [RS:0;3b3375d1a925:38409 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C38409%2C1733160278146:(num 1733160353349) 2024-12-02T17:26:13,624 DEBUG [RS:0;3b3375d1a925:38409 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:13,624 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:26:13,624 INFO [RS:0;3b3375d1a925:38409 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:26:13,624 INFO [RS:0;3b3375d1a925:38409 {}] hbase.ChoreService(370): Chore service for: regionserver/3b3375d1a925:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-02T17:26:13,624 INFO [RS:0;3b3375d1a925:38409 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:26:13,624 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:26:13,625 INFO [RS:0;3b3375d1a925:38409 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:38409 2024-12-02T17:26:13,629 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/3b3375d1a925,38409,1733160278146 2024-12-02T17:26:13,629 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:26:13,629 INFO [RS:0;3b3375d1a925:38409 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:26:13,630 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [3b3375d1a925,38409,1733160278146] 2024-12-02T17:26:13,631 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/3b3375d1a925,38409,1733160278146 already deleted, retry=false 2024-12-02T17:26:13,632 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 3b3375d1a925,38409,1733160278146 expired; onlineServers=0 2024-12-02T17:26:13,632 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '3b3375d1a925,37801,1733160277462' ***** 2024-12-02T17:26:13,632 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T17:26:13,632 INFO [M:0;3b3375d1a925:37801 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:26:13,632 INFO [M:0;3b3375d1a925:37801 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:26:13,632 DEBUG [M:0;3b3375d1a925:37801 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T17:26:13,633 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T17:26:13,633 DEBUG [M:0;3b3375d1a925:37801 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T17:26:13,633 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160279573 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160279573,5,FailOnTimeoutGroup] 2024-12-02T17:26:13,633 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160279578 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160279578,5,FailOnTimeoutGroup] 2024-12-02T17:26:13,633 INFO [M:0;3b3375d1a925:37801 {}] hbase.ChoreService(370): Chore service for: master/3b3375d1a925:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T17:26:13,633 INFO [M:0;3b3375d1a925:37801 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:26:13,633 DEBUG [M:0;3b3375d1a925:37801 {}] master.HMaster(1795): Stopping service threads 2024-12-02T17:26:13,633 INFO [M:0;3b3375d1a925:37801 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T17:26:13,633 INFO [M:0;3b3375d1a925:37801 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:26:13,634 INFO [M:0;3b3375d1a925:37801 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T17:26:13,634 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T17:26:13,634 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T17:26:13,635 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:13,635 DEBUG [M:0;3b3375d1a925:37801 {}] zookeeper.ZKUtil(347): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T17:26:13,635 WARN [M:0;3b3375d1a925:37801 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T17:26:13,636 INFO [M:0;3b3375d1a925:37801 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/.lastflushedseqids 2024-12-02T17:26:13,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741854_1030 (size=130) 2024-12-02T17:26:13,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741854_1030 (size=130) 2024-12-02T17:26:13,651 INFO [M:0;3b3375d1a925:37801 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T17:26:13,651 INFO [M:0;3b3375d1a925:37801 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T17:26:13,651 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:26:13,651 INFO [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:13,652 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:13,652 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:26:13,652 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:13,652 INFO [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=23.02 KB heapSize=29.20 KB 2024-12-02T17:26:13,676 DEBUG [M:0;3b3375d1a925:37801 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e3c01f4776c543648e31ebf5726c5c19 is 82, key is hbase:meta,,1/info:regioninfo/1733160280499/Put/seqid=0 2024-12-02T17:26:13,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741855_1031 (size=5672) 2024-12-02T17:26:13,682 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741855_1031 (size=5672) 2024-12-02T17:26:13,682 INFO [M:0;3b3375d1a925:37801 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=59 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e3c01f4776c543648e31ebf5726c5c19 2024-12-02T17:26:13,704 DEBUG [M:0;3b3375d1a925:37801 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ce65880ee608455e96d366f82826fc7c is 766, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733160281769/Put/seqid=0 2024-12-02T17:26:13,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741856_1032 (size=6247) 2024-12-02T17:26:13,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741856_1032 (size=6247) 2024-12-02T17:26:13,710 INFO [M:0;3b3375d1a925:37801 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=22.42 KB at sequenceid=59 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ce65880ee608455e96d366f82826fc7c 2024-12-02T17:26:13,717 INFO [M:0;3b3375d1a925:37801 {}] regionserver.StoreFileReader(518): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for ce65880ee608455e96d366f82826fc7c 2024-12-02T17:26:13,731 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:13,731 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38409-0x1009c051d0c0001, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:13,732 INFO [RS:0;3b3375d1a925:38409 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:26:13,732 INFO [RS:0;3b3375d1a925:38409 {}] regionserver.HRegionServer(1031): Exiting; stopping=3b3375d1a925,38409,1733160278146; zookeeper connection closed. 2024-12-02T17:26:13,732 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@1ddf13f3 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@1ddf13f3 2024-12-02T17:26:13,733 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T17:26:13,738 DEBUG [M:0;3b3375d1a925:37801 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/a8abf171c2d84570b792012c0290ff27 is 69, key is 3b3375d1a925,38409,1733160278146/rs:state/1733160279646/Put/seqid=0 2024-12-02T17:26:13,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741857_1033 (size=5156) 2024-12-02T17:26:13,744 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741857_1033 (size=5156) 2024-12-02T17:26:13,744 INFO [M:0;3b3375d1a925:37801 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=59 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/a8abf171c2d84570b792012c0290ff27 2024-12-02T17:26:13,765 DEBUG [M:0;3b3375d1a925:37801 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/c608a61cbf9b48db9d5c79e5a74619cd is 52, key is load_balancer_on/state:d/1733160280844/Put/seqid=0 2024-12-02T17:26:13,771 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741858_1034 (size=5056) 2024-12-02T17:26:13,771 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741858_1034 (size=5056) 2024-12-02T17:26:13,771 INFO [M:0;3b3375d1a925:37801 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=59 (bloomFilter=true), to=hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/c608a61cbf9b48db9d5c79e5a74619cd 2024-12-02T17:26:13,779 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e3c01f4776c543648e31ebf5726c5c19 as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/e3c01f4776c543648e31ebf5726c5c19 2024-12-02T17:26:13,786 INFO [M:0;3b3375d1a925:37801 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/e3c01f4776c543648e31ebf5726c5c19, entries=8, sequenceid=59, filesize=5.5 K 2024-12-02T17:26:13,788 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ce65880ee608455e96d366f82826fc7c as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ce65880ee608455e96d366f82826fc7c 2024-12-02T17:26:13,792 INFO [regionserver/3b3375d1a925:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:26:13,795 INFO [M:0;3b3375d1a925:37801 {}] regionserver.StoreFileReader(518): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for ce65880ee608455e96d366f82826fc7c 2024-12-02T17:26:13,795 INFO [M:0;3b3375d1a925:37801 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ce65880ee608455e96d366f82826fc7c, entries=6, sequenceid=59, filesize=6.1 K 2024-12-02T17:26:13,797 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/a8abf171c2d84570b792012c0290ff27 as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/a8abf171c2d84570b792012c0290ff27 2024-12-02T17:26:13,803 INFO [M:0;3b3375d1a925:37801 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/a8abf171c2d84570b792012c0290ff27, entries=1, sequenceid=59, filesize=5.0 K 2024-12-02T17:26:13,804 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/c608a61cbf9b48db9d5c79e5a74619cd as hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/c608a61cbf9b48db9d5c79e5a74619cd 2024-12-02T17:26:13,811 INFO [M:0;3b3375d1a925:37801 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/c608a61cbf9b48db9d5c79e5a74619cd, entries=1, sequenceid=59, filesize=4.9 K 2024-12-02T17:26:13,812 INFO [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(3140): Finished flush of dataSize ~23.02 KB/23576, heapSize ~29.14 KB/29840, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 160ms, sequenceid=59, compaction requested=false 2024-12-02T17:26:13,814 INFO [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:13,814 DEBUG [M:0;3b3375d1a925:37801 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160373651Disabling compacts and flushes for region at 1733160373651Disabling writes for close at 1733160373652 (+1 ms)Obtaining lock to block concurrent updates at 1733160373652Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733160373652Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=23576, getHeapSize=29840, getOffHeapSize=0, getCellsCount=70 at 1733160373653 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733160373653Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733160373653Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733160373676 (+23 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733160373676Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733160373688 (+12 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733160373703 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733160373703Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733160373717 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733160373737 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733160373737Flushing 1595e783b53d99cd5eef43b6debb2682/state: creating writer at 1733160373750 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: appending metadata at 1733160373765 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: closing flushed file at 1733160373765Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6057bb25: reopening flushed file at 1733160373778 (+13 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@18ddaed4: reopening flushed file at 1733160373786 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@13a5291b: reopening flushed file at 1733160373795 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3b92e31: reopening flushed file at 1733160373803 (+8 ms)Finished flush of dataSize ~23.02 KB/23576, heapSize ~29.14 KB/29840, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 160ms, sequenceid=59, compaction requested=false at 1733160373812 (+9 ms)Writing region close event to WAL at 1733160373814 (+2 ms)Closed at 1733160373814 2024-12-02T17:26:13,815 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,815 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,815 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,815 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,815 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:13,818 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33621 is added to blk_1073741830_1006 (size=27973) 2024-12-02T17:26:13,818 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44647 is added to blk_1073741830_1006 (size=27973) 2024-12-02T17:26:13,819 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:26:13,819 INFO [M:0;3b3375d1a925:37801 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T17:26:13,819 INFO [M:0;3b3375d1a925:37801 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:37801 2024-12-02T17:26:13,820 INFO [M:0;3b3375d1a925:37801 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:26:13,922 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:13,922 INFO [M:0;3b3375d1a925:37801 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:26:13,922 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:37801-0x1009c051d0c0000, quorum=127.0.0.1:65020, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:13,926 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1bf97579{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:13,928 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@22b88bcb{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:13,928 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:13,928 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2d48d695{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:13,929 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@11effdcd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:13,932 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:13,932 WARN [BP-25736929-172.17.0.3-1733160274475 heartbeating to localhost/127.0.0.1:38211 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:13,932 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:13,932 WARN [BP-25736929-172.17.0.3-1733160274475 heartbeating to localhost/127.0.0.1:38211 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-25736929-172.17.0.3-1733160274475 (Datanode Uuid 52b09fe7-8d70-4f4b-88b2-9b7bb07d5e81) service to localhost/127.0.0.1:38211 2024-12-02T17:26:13,933 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/data/data3/current/BP-25736929-172.17.0.3-1733160274475 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:13,933 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/data/data4/current/BP-25736929-172.17.0.3-1733160274475 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:13,934 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:13,939 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7b07d1ba{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:13,939 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@43e0a762{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:13,939 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:13,940 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@371e191c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:13,940 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@28778f0f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:13,941 WARN [BP-25736929-172.17.0.3-1733160274475 heartbeating to localhost/127.0.0.1:38211 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:13,941 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:13,941 WARN [BP-25736929-172.17.0.3-1733160274475 heartbeating to localhost/127.0.0.1:38211 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-25736929-172.17.0.3-1733160274475 (Datanode Uuid 8d44a392-54db-45ff-91f0-10718f9af9bc) service to localhost/127.0.0.1:38211 2024-12-02T17:26:13,942 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:13,942 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/data/data1/current/BP-25736929-172.17.0.3-1733160274475 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:13,942 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/cluster_96f71585-1d21-1981-3a01-23721af153a1/data/data2/current/BP-25736929-172.17.0.3-1733160274475 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:13,943 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:13,954 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@735fa16a{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:26:13,955 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@6c26a5a3{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:13,955 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:13,955 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@70be1389{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:13,956 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@ddc8467{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:13,967 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T17:26:14,008 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T17:26:14,021 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testSlowSyncLogRolling Thread=77 (was 12) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-2 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:38211 from jenkins.hfs.0 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: sync.1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: sync.2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-3-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: sync.2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:38211 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: sync.2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:38211 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Idle-Rpc-Conn-Sweeper-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Time-limited test.named-queue-events-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-4-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Monitor thread for TaskMonitor java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: sync.0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: master/3b3375d1a925:0:becomeActiveMaster-MemStoreChunkPool Statistics java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:38211 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: SSL Certificates Store Monitor java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Potentially hanging thread: ForkJoinPool-2-worker-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: weak-ref-cleaner-strictcontextstorage java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//io.opentelemetry.context.StrictContextStorage$PendingScopes.run(StrictContextStorage.java:269) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: sync.4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: sync.0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: HMaster-EventLoopGroup-1-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-5-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-1-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-3-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-3 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: sync.4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-1 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-2-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-3-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-4 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:38211 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: sync.1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: nioEventLoopGroup-4-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: master/3b3375d1a925:0:becomeActiveMaster-MemStoreChunkPool Statistics java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.0@localhost:38211 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: sync.3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: org.apache.hadoop.hdfs.PeerCache@7a3d4a7b java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HBase-Metrics2-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-3-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RpcClient-timer-pool-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-3-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: sync.1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: nioEventLoopGroup-2-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-1-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Timer for 'HBase' metrics system java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: nioEventLoopGroup-5-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async-Client-Retry-Timer-pool-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: SnapshotHandlerChoreCleaner java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:38211 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-4-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:38211 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-2-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-5-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: sync.4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: SessionTracker java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Potentially hanging thread: sync.0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: sync.3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: sync.3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.takeSyncRequest(FSHLog.java:426) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$SyncRunner.run(FSHLog.java:441) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-3-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: regionserver/3b3375d1a925:0.procedureResultReporter java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) - Thread LEAK? -, OpenFileDescriptor=402 (was 287) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=266 (was 419), ProcessCount=11 (was 11), AvailableMemoryMB=2320 (was 2987) 2024-12-02T17:26:14,028 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnDatanodeDeath Thread=78, OpenFileDescriptor=402, MaxFileDescriptor=1048576, SystemLoadAverage=266, ProcessCount=11, AvailableMemoryMB=2319 2024-12-02T17:26:14,029 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T17:26:14,029 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.log.dir so I do NOT create it in target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f 2024-12-02T17:26:14,029 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/229a7746-594f-9490-986f-b9902f00de0b/hadoop.tmp.dir so I do NOT create it in target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f 2024-12-02T17:26:14,029 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850, deleteOnExit=true 2024-12-02T17:26:14,029 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-02T17:26:14,029 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/test.cache.data in system properties and HBase conf 2024-12-02T17:26:14,029 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T17:26:14,029 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.log.dir in system properties and HBase conf 2024-12-02T17:26:14,030 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T17:26:14,030 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T17:26:14,030 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T17:26:14,030 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T17:26:14,030 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:26:14,030 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:26:14,030 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T17:26:14,030 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/nfs.dump.dir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/java.io.tmpdir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T17:26:14,031 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T17:26:14,045 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:26:14,118 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:14,125 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:14,126 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:14,126 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:14,126 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:26:14,127 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:14,127 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6a69944b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:14,128 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1ae773f8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:14,244 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1dc8c743{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/java.io.tmpdir/jetty-localhost-33351-hadoop-hdfs-3_4_1-tests_jar-_-any-1869079655342992664/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:26:14,245 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@6c4717a0{HTTP/1.1, (http/1.1)}{localhost:33351} 2024-12-02T17:26:14,245 INFO [Time-limited test {}] server.Server(415): Started @101917ms 2024-12-02T17:26:14,258 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:26:14,327 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:14,331 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:14,332 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:14,332 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:14,332 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:26:14,332 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@670e4080{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:14,333 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7247ee1d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:14,450 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6a249094{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/java.io.tmpdir/jetty-localhost-43545-hadoop-hdfs-3_4_1-tests_jar-_-any-7069446176596330557/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:14,450 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@35445cbd{HTTP/1.1, (http/1.1)}{localhost:43545} 2024-12-02T17:26:14,450 INFO [Time-limited test {}] server.Server(415): Started @102122ms 2024-12-02T17:26:14,452 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:14,485 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:14,489 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:14,490 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:14,490 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:14,490 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:26:14,490 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@18f27499{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:14,491 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@64e5ce98{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:14,538 WARN [Thread-437 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/data/data1/current/BP-729187193-172.17.0.3-1733160374064/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:14,538 WARN [Thread-438 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/data/data2/current/BP-729187193-172.17.0.3-1733160374064/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:14,558 WARN [Thread-416 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:14,562 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x157a68cb5cc44d1e with lease ID 0x2ac96f6f8b0cdafc: Processing first storage report for DS-112e0df4-c337-4ecc-928b-ebeedd9d501b from datanode DatanodeRegistration(127.0.0.1:35293, datanodeUuid=b4a9e621-0374-407c-b999-17e2ccfe517a, infoPort=42567, infoSecurePort=0, ipcPort=40637, storageInfo=lv=-57;cid=testClusterID;nsid=1619137357;c=1733160374064) 2024-12-02T17:26:14,562 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x157a68cb5cc44d1e with lease ID 0x2ac96f6f8b0cdafc: from storage DS-112e0df4-c337-4ecc-928b-ebeedd9d501b node DatanodeRegistration(127.0.0.1:35293, datanodeUuid=b4a9e621-0374-407c-b999-17e2ccfe517a, infoPort=42567, infoSecurePort=0, ipcPort=40637, storageInfo=lv=-57;cid=testClusterID;nsid=1619137357;c=1733160374064), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T17:26:14,562 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x157a68cb5cc44d1e with lease ID 0x2ac96f6f8b0cdafc: Processing first storage report for DS-3319fd45-32af-418e-bad7-f5e9317e61ed from datanode DatanodeRegistration(127.0.0.1:35293, datanodeUuid=b4a9e621-0374-407c-b999-17e2ccfe517a, infoPort=42567, infoSecurePort=0, ipcPort=40637, storageInfo=lv=-57;cid=testClusterID;nsid=1619137357;c=1733160374064) 2024-12-02T17:26:14,562 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x157a68cb5cc44d1e with lease ID 0x2ac96f6f8b0cdafc: from storage DS-3319fd45-32af-418e-bad7-f5e9317e61ed node DatanodeRegistration(127.0.0.1:35293, datanodeUuid=b4a9e621-0374-407c-b999-17e2ccfe517a, infoPort=42567, infoSecurePort=0, ipcPort=40637, storageInfo=lv=-57;cid=testClusterID;nsid=1619137357;c=1733160374064), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:14,609 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@254b7745{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/java.io.tmpdir/jetty-localhost-42295-hadoop-hdfs-3_4_1-tests_jar-_-any-17657175641765231626/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:14,609 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@86e294e{HTTP/1.1, (http/1.1)}{localhost:42295} 2024-12-02T17:26:14,610 INFO [Time-limited test {}] server.Server(415): Started @102282ms 2024-12-02T17:26:14,611 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:14,701 WARN [Thread-463 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/data/data3/current/BP-729187193-172.17.0.3-1733160374064/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:14,701 WARN [Thread-464 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/data/data4/current/BP-729187193-172.17.0.3-1733160374064/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:14,722 WARN [Thread-452 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:14,725 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe0395153a54bafa6 with lease ID 0x2ac96f6f8b0cdafd: Processing first storage report for DS-82a98fee-258e-4743-8bf5-4a8cda3fc114 from datanode DatanodeRegistration(127.0.0.1:34987, datanodeUuid=ce042a7b-d154-4ddf-aa79-94dc7f731855, infoPort=46575, infoSecurePort=0, ipcPort=40059, storageInfo=lv=-57;cid=testClusterID;nsid=1619137357;c=1733160374064) 2024-12-02T17:26:14,726 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe0395153a54bafa6 with lease ID 0x2ac96f6f8b0cdafd: from storage DS-82a98fee-258e-4743-8bf5-4a8cda3fc114 node DatanodeRegistration(127.0.0.1:34987, datanodeUuid=ce042a7b-d154-4ddf-aa79-94dc7f731855, infoPort=46575, infoSecurePort=0, ipcPort=40059, storageInfo=lv=-57;cid=testClusterID;nsid=1619137357;c=1733160374064), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:14,726 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe0395153a54bafa6 with lease ID 0x2ac96f6f8b0cdafd: Processing first storage report for DS-2545e434-7ea2-4cb4-8dad-c05936aac110 from datanode DatanodeRegistration(127.0.0.1:34987, datanodeUuid=ce042a7b-d154-4ddf-aa79-94dc7f731855, infoPort=46575, infoSecurePort=0, ipcPort=40059, storageInfo=lv=-57;cid=testClusterID;nsid=1619137357;c=1733160374064) 2024-12-02T17:26:14,726 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe0395153a54bafa6 with lease ID 0x2ac96f6f8b0cdafd: from storage DS-2545e434-7ea2-4cb4-8dad-c05936aac110 node DatanodeRegistration(127.0.0.1:34987, datanodeUuid=ce042a7b-d154-4ddf-aa79-94dc7f731855, infoPort=46575, infoSecurePort=0, ipcPort=40059, storageInfo=lv=-57;cid=testClusterID;nsid=1619137357;c=1733160374064), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:14,739 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f 2024-12-02T17:26:14,742 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/zookeeper_0, clientPort=63380, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T17:26:14,744 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=63380 2024-12-02T17:26:14,744 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:14,746 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:14,760 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:26:14,760 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:26:14,762 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869 with version=8 2024-12-02T17:26:14,762 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase-staging 2024-12-02T17:26:14,764 INFO [Time-limited test {}] client.ConnectionUtils(128): master/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:26:14,764 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:14,764 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:14,764 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:26:14,764 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:14,764 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:26:14,764 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T17:26:14,765 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:26:14,765 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:41757 2024-12-02T17:26:14,767 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:41757 connecting to ZooKeeper ensemble=127.0.0.1:63380 2024-12-02T17:26:14,773 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:417570x0, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:26:14,774 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:41757-0x1009c069c740000 connected 2024-12-02T17:26:14,793 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:14,794 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:14,797 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:14,797 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869, hbase.cluster.distributed=false 2024-12-02T17:26:14,798 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:26:14,799 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41757 2024-12-02T17:26:14,799 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41757 2024-12-02T17:26:14,799 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41757 2024-12-02T17:26:14,800 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41757 2024-12-02T17:26:14,800 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41757 2024-12-02T17:26:14,817 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:26:14,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:14,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:14,817 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:26:14,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:14,818 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:26:14,818 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T17:26:14,818 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:26:14,818 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:33327 2024-12-02T17:26:14,820 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:33327 connecting to ZooKeeper ensemble=127.0.0.1:63380 2024-12-02T17:26:14,820 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:14,823 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:14,827 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:333270x0, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:26:14,827 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:33327-0x1009c069c740001 connected 2024-12-02T17:26:14,827 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:14,828 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T17:26:14,831 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T17:26:14,831 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T17:26:14,833 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:26:14,833 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=33327 2024-12-02T17:26:14,833 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=33327 2024-12-02T17:26:14,833 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=33327 2024-12-02T17:26:14,838 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=33327 2024-12-02T17:26:14,839 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=33327 2024-12-02T17:26:14,852 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;3b3375d1a925:41757 2024-12-02T17:26:14,853 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/3b3375d1a925,41757,1733160374764 2024-12-02T17:26:14,854 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:26:14,855 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:26:14,855 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/3b3375d1a925,41757,1733160374764 2024-12-02T17:26:14,857 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T17:26:14,857 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:14,857 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:14,857 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T17:26:14,858 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/3b3375d1a925,41757,1733160374764 from backup master directory 2024-12-02T17:26:14,859 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/3b3375d1a925,41757,1733160374764 2024-12-02T17:26:14,859 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:26:14,859 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:26:14,860 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=3b3375d1a925,41757,1733160374764 2024-12-02T17:26:14,860 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:26:14,866 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/hbase.id] with ID: 27ce32d6-3622-4d11-a307-466ee9df1d9f 2024-12-02T17:26:14,866 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/.tmp/hbase.id 2024-12-02T17:26:14,875 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:26:14,876 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:26:14,877 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/.tmp/hbase.id]:[hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/hbase.id] 2024-12-02T17:26:14,892 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:14,892 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T17:26:14,894 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-12-02T17:26:14,896 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:14,896 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:14,912 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:26:14,913 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:26:14,914 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:26:14,915 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T17:26:14,915 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:26:14,935 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:26:14,935 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:26:14,936 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store 2024-12-02T17:26:14,945 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:26:14,946 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:26:14,947 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:14,948 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:26:14,948 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:14,948 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:14,948 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:26:14,948 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:14,948 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:14,948 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160374948Disabling compacts and flushes for region at 1733160374948Disabling writes for close at 1733160374948Writing region close event to WAL at 1733160374948Closed at 1733160374948 2024-12-02T17:26:14,949 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/.initializing 2024-12-02T17:26:14,949 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/WALs/3b3375d1a925,41757,1733160374764 2024-12-02T17:26:14,953 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C41757%2C1733160374764, suffix=, logDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/WALs/3b3375d1a925,41757,1733160374764, archiveDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/oldWALs, maxLogs=10 2024-12-02T17:26:14,953 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C41757%2C1733160374764.1733160374953 2024-12-02T17:26:14,961 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/WALs/3b3375d1a925,41757,1733160374764/3b3375d1a925%2C41757%2C1733160374764.1733160374953 2024-12-02T17:26:14,963 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46575:46575),(127.0.0.1/127.0.0.1:42567:42567)] 2024-12-02T17:26:14,966 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:26:14,967 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:14,967 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,967 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,969 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,971 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T17:26:14,971 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:14,971 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:14,972 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,973 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T17:26:14,973 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:14,974 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:26:14,974 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,976 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T17:26:14,977 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:14,977 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:26:14,977 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,979 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T17:26:14,979 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:14,979 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:26:14,980 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,981 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,981 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,983 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,983 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,984 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T17:26:14,985 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:14,989 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:26:14,989 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=775277, jitterRate=-0.014185145497322083}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T17:26:14,991 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733160374967Initializing all the Stores at 1733160374968 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160374968Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160374969 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160374969Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160374969Cleaning up temporary data from old regions at 1733160374983 (+14 ms)Region opened successfully at 1733160374990 (+7 ms) 2024-12-02T17:26:14,991 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T17:26:14,998 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5a3d055d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:26:14,999 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T17:26:14,999 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T17:26:14,999 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T17:26:14,999 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T17:26:15,000 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T17:26:15,000 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-02T17:26:15,000 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T17:26:15,003 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T17:26:15,004 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T17:26:15,006 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T17:26:15,006 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T17:26:15,007 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T17:26:15,008 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T17:26:15,009 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T17:26:15,010 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T17:26:15,011 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T17:26:15,012 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T17:26:15,013 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T17:26:15,016 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T17:26:15,017 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T17:26:15,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:15,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:15,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,020 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=3b3375d1a925,41757,1733160374764, sessionid=0x1009c069c740000, setting cluster-up flag (Was=false) 2024-12-02T17:26:15,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,028 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T17:26:15,029 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,41757,1733160374764 2024-12-02T17:26:15,033 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,033 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,037 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T17:26:15,038 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,41757,1733160374764 2024-12-02T17:26:15,040 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T17:26:15,041 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T17:26:15,042 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T17:26:15,042 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T17:26:15,042 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 3b3375d1a925,41757,1733160374764 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T17:26:15,043 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:26:15,044 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:26:15,044 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:26:15,044 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:26:15,044 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/3b3375d1a925:0, corePoolSize=10, maxPoolSize=10 2024-12-02T17:26:15,044 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,044 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:26:15,044 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,045 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733160405045 2024-12-02T17:26:15,045 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T17:26:15,046 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T17:26:15,046 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(746): ClusterId : 27ce32d6-3622-4d11-a307-466ee9df1d9f 2024-12-02T17:26:15,046 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T17:26:15,046 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T17:26:15,046 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T17:26:15,046 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T17:26:15,046 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T17:26:15,046 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,046 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:26:15,046 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T17:26:15,046 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T17:26:15,047 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T17:26:15,047 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T17:26:15,047 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T17:26:15,047 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T17:26:15,047 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160375047,5,FailOnTimeoutGroup] 2024-12-02T17:26:15,047 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160375047,5,FailOnTimeoutGroup] 2024-12-02T17:26:15,048 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,048 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T17:26:15,048 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,048 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,048 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,048 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T17:26:15,048 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T17:26:15,048 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T17:26:15,050 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T17:26:15,051 DEBUG [RS:0;3b3375d1a925:33327 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6181265a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:26:15,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:26:15,056 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:26:15,057 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T17:26:15,057 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869 2024-12-02T17:26:15,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:26:15,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:26:15,070 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:15,071 DEBUG [RS:0;3b3375d1a925:33327 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;3b3375d1a925:33327 2024-12-02T17:26:15,071 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T17:26:15,071 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T17:26:15,071 DEBUG [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T17:26:15,071 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:26:15,072 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(2659): reportForDuty to master=3b3375d1a925,41757,1733160374764 with port=33327, startcode=1733160374817 2024-12-02T17:26:15,073 DEBUG [RS:0;3b3375d1a925:33327 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T17:26:15,073 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:26:15,073 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,074 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:15,074 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:26:15,076 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:26:15,077 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,077 INFO [HMaster-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:33945, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T17:26:15,077 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:15,077 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:26:15,078 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41757 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,078 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41757 {}] master.ServerManager(517): Registering regionserver=3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,079 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:26:15,079 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,080 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:15,080 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:26:15,081 DEBUG [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869 2024-12-02T17:26:15,081 DEBUG [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:35133 2024-12-02T17:26:15,081 DEBUG [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T17:26:15,082 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:26:15,082 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,083 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:26:15,083 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:15,083 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:26:15,084 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740 2024-12-02T17:26:15,084 DEBUG [RS:0;3b3375d1a925:33327 {}] zookeeper.ZKUtil(111): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,084 WARN [RS:0;3b3375d1a925:33327 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:26:15,084 INFO [RS:0;3b3375d1a925:33327 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:26:15,084 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740 2024-12-02T17:26:15,085 DEBUG [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/WALs/3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,085 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [3b3375d1a925,33327,1733160374817] 2024-12-02T17:26:15,086 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:26:15,086 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:26:15,087 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:26:15,089 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:26:15,093 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T17:26:15,093 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:26:15,094 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=878289, jitterRate=0.11680243909358978}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:26:15,096 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733160375070Initializing all the Stores at 1733160375071 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160375071Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160375071Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160375071Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160375071Cleaning up temporary data from old regions at 1733160375086 (+15 ms)Region opened successfully at 1733160375095 (+9 ms) 2024-12-02T17:26:15,096 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:26:15,096 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:26:15,096 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:26:15,096 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:26:15,096 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:26:15,097 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:26:15,097 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160375096Disabling compacts and flushes for region at 1733160375096Disabling writes for close at 1733160375096Writing region close event to WAL at 1733160375097 (+1 ms)Closed at 1733160375097 2024-12-02T17:26:15,097 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T17:26:15,098 INFO [RS:0;3b3375d1a925:33327 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:26:15,098 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,098 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T17:26:15,099 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:26:15,099 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T17:26:15,099 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T17:26:15,099 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,099 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T17:26:15,099 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,099 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,099 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:26:15,100 DEBUG [RS:0;3b3375d1a925:33327 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:26:15,101 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,101 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,101 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,101 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,101 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,101 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,33327,1733160374817-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:26:15,101 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:26:15,103 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T17:26:15,120 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T17:26:15,120 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,33327,1733160374817-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,121 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,121 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.Replication(171): 3b3375d1a925,33327,1733160374817 started 2024-12-02T17:26:15,136 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,137 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1482): Serving as 3b3375d1a925,33327,1733160374817, RpcServer on 3b3375d1a925/172.17.0.3:33327, sessionid=0x1009c069c740001 2024-12-02T17:26:15,137 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T17:26:15,137 DEBUG [RS:0;3b3375d1a925:33327 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,137 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,33327,1733160374817' 2024-12-02T17:26:15,137 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T17:26:15,138 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T17:26:15,139 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T17:26:15,139 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T17:26:15,139 DEBUG [RS:0;3b3375d1a925:33327 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,139 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,33327,1733160374817' 2024-12-02T17:26:15,139 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T17:26:15,139 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T17:26:15,140 DEBUG [RS:0;3b3375d1a925:33327 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T17:26:15,140 INFO [RS:0;3b3375d1a925:33327 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T17:26:15,140 INFO [RS:0;3b3375d1a925:33327 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T17:26:15,243 INFO [RS:0;3b3375d1a925:33327 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C33327%2C1733160374817, suffix=, logDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/WALs/3b3375d1a925,33327,1733160374817, archiveDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/oldWALs, maxLogs=32 2024-12-02T17:26:15,245 INFO [RS:0;3b3375d1a925:33327 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C33327%2C1733160374817.1733160375245 2024-12-02T17:26:15,252 INFO [RS:0;3b3375d1a925:33327 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/WALs/3b3375d1a925,33327,1733160374817/3b3375d1a925%2C33327%2C1733160374817.1733160375245 2024-12-02T17:26:15,253 DEBUG [RS:0;3b3375d1a925:33327 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46575:46575),(127.0.0.1/127.0.0.1:42567:42567)] 2024-12-02T17:26:15,253 WARN [3b3375d1a925:41757 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-02T17:26:15,503 DEBUG [3b3375d1a925:41757 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T17:26:15,504 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,507 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,33327,1733160374817, state=OPENING 2024-12-02T17:26:15,509 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T17:26:15,510 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,510 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,512 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:26:15,512 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,33327,1733160374817}] 2024-12-02T17:26:15,512 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:26:15,513 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:26:15,667 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T17:26:15,670 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-6-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:44885, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T17:26:15,675 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T17:26:15,675 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:26:15,678 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C33327%2C1733160374817.meta, suffix=.meta, logDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/WALs/3b3375d1a925,33327,1733160374817, archiveDir=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/oldWALs, maxLogs=32 2024-12-02T17:26:15,681 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C33327%2C1733160374817.meta.1733160375680.meta 2024-12-02T17:26:15,689 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/WALs/3b3375d1a925,33327,1733160374817/3b3375d1a925%2C33327%2C1733160374817.meta.1733160375680.meta 2024-12-02T17:26:15,693 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:42567:42567),(127.0.0.1/127.0.0.1:46575:46575)] 2024-12-02T17:26:15,698 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:26:15,698 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T17:26:15,698 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T17:26:15,699 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T17:26:15,699 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T17:26:15,699 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:15,699 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T17:26:15,699 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T17:26:15,701 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:26:15,703 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:26:15,703 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,703 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:15,704 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:26:15,705 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:26:15,705 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,705 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:15,706 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:26:15,707 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:26:15,707 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,709 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:15,709 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:26:15,710 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:26:15,710 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:15,710 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:15,711 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:26:15,712 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740 2024-12-02T17:26:15,714 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740 2024-12-02T17:26:15,716 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:26:15,716 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:26:15,717 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:26:15,719 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:26:15,721 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=838301, jitterRate=0.06595569849014282}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:26:15,721 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T17:26:15,722 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733160375699Writing region info on filesystem at 1733160375699Initializing all the Stores at 1733160375701 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160375701Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160375701Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160375701Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160375701Cleaning up temporary data from old regions at 1733160375716 (+15 ms)Running coprocessor post-open hooks at 1733160375721 (+5 ms)Region opened successfully at 1733160375722 (+1 ms) 2024-12-02T17:26:15,724 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733160375666 2024-12-02T17:26:15,728 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T17:26:15,728 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T17:26:15,729 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,731 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,33327,1733160374817, state=OPEN 2024-12-02T17:26:15,738 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:26:15,738 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:26:15,738 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,738 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:26:15,738 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:26:15,742 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T17:26:15,742 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,33327,1733160374817 in 226 msec 2024-12-02T17:26:15,746 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T17:26:15,746 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 643 msec 2024-12-02T17:26:15,748 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:26:15,748 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T17:26:15,750 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:26:15,750 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,33327,1733160374817, seqNum=-1] 2024-12-02T17:26:15,750 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:26:15,752 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-6-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51065, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:26:15,760 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 717 msec 2024-12-02T17:26:15,760 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733160375760, completionTime=-1 2024-12-02T17:26:15,761 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T17:26:15,761 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T17:26:15,763 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=1 2024-12-02T17:26:15,763 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733160435763 2024-12-02T17:26:15,763 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733160495763 2024-12-02T17:26:15,764 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-12-02T17:26:15,764 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,41757,1733160374764-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,764 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,41757,1733160374764-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,764 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,41757,1733160374764-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,764 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-3b3375d1a925:41757, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,764 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,767 DEBUG [master/3b3375d1a925:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T17:26:15,768 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,774 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.914sec 2024-12-02T17:26:15,774 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T17:26:15,774 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T17:26:15,774 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T17:26:15,774 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T17:26:15,775 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T17:26:15,775 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,41757,1733160374764-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:26:15,775 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,41757,1733160374764-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T17:26:15,778 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T17:26:15,778 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T17:26:15,778 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,41757,1733160374764-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:15,846 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2fa3cb60, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:26:15,846 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 3b3375d1a925,41757,-1 for getting cluster id 2024-12-02T17:26:15,846 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T17:26:15,848 DEBUG [HMaster-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '27ce32d6-3622-4d11-a307-466ee9df1d9f' 2024-12-02T17:26:15,849 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T17:26:15,849 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "27ce32d6-3622-4d11-a307-466ee9df1d9f" 2024-12-02T17:26:15,849 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7d780044, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:26:15,849 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [3b3375d1a925,41757,-1] 2024-12-02T17:26:15,850 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T17:26:15,851 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:15,852 INFO [HMaster-EventLoopGroup-5-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:36314, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T17:26:15,853 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1f14acfa, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:26:15,854 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:26:15,855 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,33327,1733160374817, seqNum=-1] 2024-12-02T17:26:15,855 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:26:15,857 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-6-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:50216, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:26:15,858 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=3b3375d1a925,41757,1733160374764 2024-12-02T17:26:15,859 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:15,862 INFO [Time-limited test {}] master.MasterRpcServices(567): Client=null/null set balanceSwitch=false 2024-12-02T17:26:15,862 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T17:26:15,862 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:26:15,862 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at org.apache.hadoop.hbase.regionserver.wal.TestLogRolling.testLogRollOnDatanodeDeath(TestLogRolling.java:201) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:15,863 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:15,863 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:15,863 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T17:26:15,863 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T17:26:15,863 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1771574582, stopped=false 2024-12-02T17:26:15,863 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=3b3375d1a925,41757,1733160374764 2024-12-02T17:26:15,865 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:15,865 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,865 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:26:15,865 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:15,865 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:15,865 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:26:15,866 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:15,866 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at org.apache.hadoop.hbase.regionserver.wal.TestLogRolling.testLogRollOnDatanodeDeath(TestLogRolling.java:201) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:15,866 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:15,866 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:15,866 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(878): Closing user regions 2024-12-02T17:26:15,866 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '3b3375d1a925,33327,1733160374817' ***** 2024-12-02T17:26:15,867 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T17:26:15,868 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(959): stopping server 3b3375d1a925,33327,1733160374817 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;3b3375d1a925:33327. 2024-12-02T17:26:15,868 DEBUG [RS:0;3b3375d1a925:33327 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:15,868 DEBUG [RS:0;3b3375d1a925:33327 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T17:26:15,868 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T17:26:15,869 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-02T17:26:15,869 DEBUG [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-02T17:26:15,869 DEBUG [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-02T17:26:15,869 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:26:15,869 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:26:15,869 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:26:15,869 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:26:15,869 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:26:15,869 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=74 B heapSize=1.22 KB 2024-12-02T17:26:15,887 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740/.tmp/ns/1351ab966c5e47e4933a7d2d1b1f998e is 43, key is default/ns:d/1733160375753/Put/seqid=0 2024-12-02T17:26:15,893 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741835_1011 (size=5153) 2024-12-02T17:26:15,893 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741835_1011 (size=5153) 2024-12-02T17:26:15,894 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740/.tmp/ns/1351ab966c5e47e4933a7d2d1b1f998e 2024-12-02T17:26:15,903 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740/.tmp/ns/1351ab966c5e47e4933a7d2d1b1f998e as hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740/ns/1351ab966c5e47e4933a7d2d1b1f998e 2024-12-02T17:26:15,910 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740/ns/1351ab966c5e47e4933a7d2d1b1f998e, entries=2, sequenceid=6, filesize=5.0 K 2024-12-02T17:26:15,911 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~74 B/74, heapSize ~464 B/464, currentSize=0 B/0 for 1588230740 in 42ms, sequenceid=6, compaction requested=false 2024-12-02T17:26:15,916 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/data/hbase/meta/1588230740/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-02T17:26:15,917 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:26:15,917 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:26:15,917 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160375869Running coprocessor pre-close hooks at 1733160375869Disabling compacts and flushes for region at 1733160375869Disabling writes for close at 1733160375869Obtaining lock to block concurrent updates at 1733160375869Preparing flush snapshotting stores in 1588230740 at 1733160375869Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=74, getHeapSize=1184, getOffHeapSize=0, getCellsCount=2 at 1733160375870 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733160375870Flushing 1588230740/ns: creating writer at 1733160375871 (+1 ms)Flushing 1588230740/ns: appending metadata at 1733160375887 (+16 ms)Flushing 1588230740/ns: closing flushed file at 1733160375887Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7a98acfc: reopening flushed file at 1733160375902 (+15 ms)Finished flush of dataSize ~74 B/74, heapSize ~464 B/464, currentSize=0 B/0 for 1588230740 in 42ms, sequenceid=6, compaction requested=false at 1733160375911 (+9 ms)Writing region close event to WAL at 1733160375912 (+1 ms)Running coprocessor post-close hooks at 1733160375917 (+5 ms)Closed at 1733160375917 2024-12-02T17:26:15,918 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T17:26:16,069 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(976): stopping server 3b3375d1a925,33327,1733160374817; all regions closed. 2024-12-02T17:26:16,070 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,070 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,070 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,070 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,070 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741834_1010 (size=1152) 2024-12-02T17:26:16,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741834_1010 (size=1152) 2024-12-02T17:26:16,076 DEBUG [RS:0;3b3375d1a925:33327 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/oldWALs 2024-12-02T17:26:16,076 INFO [RS:0;3b3375d1a925:33327 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C33327%2C1733160374817.meta:.meta(num 1733160375680) 2024-12-02T17:26:16,076 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,076 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,077 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,077 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,077 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741833_1009 (size=93) 2024-12-02T17:26:16,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741833_1009 (size=93) 2024-12-02T17:26:16,081 DEBUG [RS:0;3b3375d1a925:33327 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/oldWALs 2024-12-02T17:26:16,081 INFO [RS:0;3b3375d1a925:33327 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C33327%2C1733160374817:(num 1733160375245) 2024-12-02T17:26:16,081 DEBUG [RS:0;3b3375d1a925:33327 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:16,081 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:26:16,081 INFO [RS:0;3b3375d1a925:33327 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:26:16,082 INFO [RS:0;3b3375d1a925:33327 {}] hbase.ChoreService(370): Chore service for: regionserver/3b3375d1a925:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-02T17:26:16,082 INFO [RS:0;3b3375d1a925:33327 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:26:16,082 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:26:16,082 INFO [RS:0;3b3375d1a925:33327 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:33327 2024-12-02T17:26:16,084 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/3b3375d1a925,33327,1733160374817 2024-12-02T17:26:16,084 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:26:16,084 INFO [RS:0;3b3375d1a925:33327 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:26:16,086 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [3b3375d1a925,33327,1733160374817] 2024-12-02T17:26:16,087 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/3b3375d1a925,33327,1733160374817 already deleted, retry=false 2024-12-02T17:26:16,087 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 3b3375d1a925,33327,1733160374817 expired; onlineServers=0 2024-12-02T17:26:16,087 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '3b3375d1a925,41757,1733160374764' ***** 2024-12-02T17:26:16,087 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T17:26:16,087 INFO [M:0;3b3375d1a925:41757 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:26:16,088 INFO [M:0;3b3375d1a925:41757 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:26:16,088 DEBUG [M:0;3b3375d1a925:41757 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T17:26:16,088 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T17:26:16,088 DEBUG [M:0;3b3375d1a925:41757 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T17:26:16,088 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160375047 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160375047,5,FailOnTimeoutGroup] 2024-12-02T17:26:16,088 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160375047 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160375047,5,FailOnTimeoutGroup] 2024-12-02T17:26:16,088 INFO [M:0;3b3375d1a925:41757 {}] hbase.ChoreService(370): Chore service for: master/3b3375d1a925:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T17:26:16,088 INFO [M:0;3b3375d1a925:41757 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:26:16,088 DEBUG [M:0;3b3375d1a925:41757 {}] master.HMaster(1795): Stopping service threads 2024-12-02T17:26:16,088 INFO [M:0;3b3375d1a925:41757 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T17:26:16,088 INFO [M:0;3b3375d1a925:41757 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:26:16,088 INFO [M:0;3b3375d1a925:41757 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T17:26:16,089 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T17:26:16,089 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T17:26:16,090 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:16,090 DEBUG [M:0;3b3375d1a925:41757 {}] zookeeper.ZKUtil(347): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T17:26:16,090 WARN [M:0;3b3375d1a925:41757 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T17:26:16,090 INFO [M:0;3b3375d1a925:41757 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/.lastflushedseqids 2024-12-02T17:26:16,096 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741836_1012 (size=99) 2024-12-02T17:26:16,096 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741836_1012 (size=99) 2024-12-02T17:26:16,097 INFO [M:0;3b3375d1a925:41757 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T17:26:16,097 INFO [M:0;3b3375d1a925:41757 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T17:26:16,097 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:26:16,098 INFO [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:16,098 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:16,098 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:26:16,098 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:16,098 INFO [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=7.67 KB heapSize=11.34 KB 2024-12-02T17:26:16,117 DEBUG [M:0;3b3375d1a925:41757 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/b6708fb357bc4995a86ed74445bf55f5 is 82, key is hbase:meta,,1/info:regioninfo/1733160375729/Put/seqid=0 2024-12-02T17:26:16,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741837_1013 (size=5672) 2024-12-02T17:26:16,123 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741837_1013 (size=5672) 2024-12-02T17:26:16,123 INFO [M:0;3b3375d1a925:41757 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=29 (bloomFilter=true), to=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/b6708fb357bc4995a86ed74445bf55f5 2024-12-02T17:26:16,145 DEBUG [M:0;3b3375d1a925:41757 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/5a2fdfd28b0f4eaa81d3cca0717cc2b0 is 240, key is \x00\x00\x00\x00\x00\x00\x00\x01/proc:d/1733160375759/Put/seqid=0 2024-12-02T17:26:16,151 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741838_1014 (size=5275) 2024-12-02T17:26:16,151 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741838_1014 (size=5275) 2024-12-02T17:26:16,152 INFO [M:0;3b3375d1a925:41757 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.06 KB at sequenceid=29 (bloomFilter=true), to=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/5a2fdfd28b0f4eaa81d3cca0717cc2b0 2024-12-02T17:26:16,173 DEBUG [M:0;3b3375d1a925:41757 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/977c4687628349d590ff643bd2d5c0aa is 69, key is 3b3375d1a925,33327,1733160374817/rs:state/1733160375079/Put/seqid=0 2024-12-02T17:26:16,178 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741839_1015 (size=5156) 2024-12-02T17:26:16,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741839_1015 (size=5156) 2024-12-02T17:26:16,179 INFO [M:0;3b3375d1a925:41757 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=29 (bloomFilter=true), to=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/977c4687628349d590ff643bd2d5c0aa 2024-12-02T17:26:16,186 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:16,186 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33327-0x1009c069c740001, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:16,186 INFO [RS:0;3b3375d1a925:33327 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:26:16,186 INFO [RS:0;3b3375d1a925:33327 {}] regionserver.HRegionServer(1031): Exiting; stopping=3b3375d1a925,33327,1733160374817; zookeeper connection closed. 2024-12-02T17:26:16,187 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@126be121 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@126be121 2024-12-02T17:26:16,187 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T17:26:16,207 DEBUG [M:0;3b3375d1a925:41757 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/07dd30984b5e4dd78b0cc117f39ac8dc is 52, key is load_balancer_on/state:d/1733160375861/Put/seqid=0 2024-12-02T17:26:16,214 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741840_1016 (size=5056) 2024-12-02T17:26:16,215 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741840_1016 (size=5056) 2024-12-02T17:26:16,215 INFO [M:0;3b3375d1a925:41757 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=29 (bloomFilter=true), to=hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/07dd30984b5e4dd78b0cc117f39ac8dc 2024-12-02T17:26:16,222 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/b6708fb357bc4995a86ed74445bf55f5 as hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/b6708fb357bc4995a86ed74445bf55f5 2024-12-02T17:26:16,229 INFO [M:0;3b3375d1a925:41757 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/b6708fb357bc4995a86ed74445bf55f5, entries=8, sequenceid=29, filesize=5.5 K 2024-12-02T17:26:16,230 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/5a2fdfd28b0f4eaa81d3cca0717cc2b0 as hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/5a2fdfd28b0f4eaa81d3cca0717cc2b0 2024-12-02T17:26:16,236 INFO [M:0;3b3375d1a925:41757 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/5a2fdfd28b0f4eaa81d3cca0717cc2b0, entries=3, sequenceid=29, filesize=5.2 K 2024-12-02T17:26:16,237 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/977c4687628349d590ff643bd2d5c0aa as hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/977c4687628349d590ff643bd2d5c0aa 2024-12-02T17:26:16,244 INFO [M:0;3b3375d1a925:41757 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/977c4687628349d590ff643bd2d5c0aa, entries=1, sequenceid=29, filesize=5.0 K 2024-12-02T17:26:16,245 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/07dd30984b5e4dd78b0cc117f39ac8dc as hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/07dd30984b5e4dd78b0cc117f39ac8dc 2024-12-02T17:26:16,251 INFO [M:0;3b3375d1a925:41757 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35133/user/jenkins/test-data/b91a554b-d06b-a778-0a8c-63c306007869/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/07dd30984b5e4dd78b0cc117f39ac8dc, entries=1, sequenceid=29, filesize=4.9 K 2024-12-02T17:26:16,253 INFO [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.67 KB/7850, heapSize ~11.27 KB/11544, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 155ms, sequenceid=29, compaction requested=false 2024-12-02T17:26:16,255 INFO [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:16,255 DEBUG [M:0;3b3375d1a925:41757 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160376097Disabling compacts and flushes for region at 1733160376097Disabling writes for close at 1733160376098 (+1 ms)Obtaining lock to block concurrent updates at 1733160376098Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733160376098Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=7850, getHeapSize=11544, getOffHeapSize=0, getCellsCount=36 at 1733160376098Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733160376099 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733160376099Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733160376116 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733160376116Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733160376129 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733160376145 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733160376145Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733160376158 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733160376173 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733160376173Flushing 1595e783b53d99cd5eef43b6debb2682/state: creating writer at 1733160376185 (+12 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: appending metadata at 1733160376207 (+22 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: closing flushed file at 1733160376207Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4f61808b: reopening flushed file at 1733160376221 (+14 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@dbc4d7b: reopening flushed file at 1733160376229 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3545acca: reopening flushed file at 1733160376236 (+7 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6e698670: reopening flushed file at 1733160376244 (+8 ms)Finished flush of dataSize ~7.67 KB/7850, heapSize ~11.27 KB/11544, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 155ms, sequenceid=29, compaction requested=false at 1733160376253 (+9 ms)Writing region close event to WAL at 1733160376254 (+1 ms)Closed at 1733160376254 2024-12-02T17:26:16,255 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,255 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,256 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,256 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,256 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:16,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35293 is added to blk_1073741830_1006 (size=10311) 2024-12-02T17:26:16,259 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34987 is added to blk_1073741830_1006 (size=10311) 2024-12-02T17:26:16,259 INFO [M:0;3b3375d1a925:41757 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T17:26:16,259 INFO [M:0;3b3375d1a925:41757 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:41757 2024-12-02T17:26:16,259 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:26:16,259 INFO [M:0;3b3375d1a925:41757 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:26:16,362 INFO [M:0;3b3375d1a925:41757 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:26:16,362 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:16,362 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41757-0x1009c069c740000, quorum=127.0.0.1:63380, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:16,364 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@254b7745{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:16,365 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@86e294e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:16,365 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:16,365 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@64e5ce98{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:16,365 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@18f27499{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:16,366 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:16,366 WARN [BP-729187193-172.17.0.3-1733160374064 heartbeating to localhost/127.0.0.1:35133 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:16,366 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:16,367 WARN [BP-729187193-172.17.0.3-1733160374064 heartbeating to localhost/127.0.0.1:35133 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-729187193-172.17.0.3-1733160374064 (Datanode Uuid ce042a7b-d154-4ddf-aa79-94dc7f731855) service to localhost/127.0.0.1:35133 2024-12-02T17:26:16,367 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/data/data3/current/BP-729187193-172.17.0.3-1733160374064 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:16,368 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/data/data4/current/BP-729187193-172.17.0.3-1733160374064 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:16,368 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:16,373 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6a249094{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:16,373 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@35445cbd{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:16,373 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:16,373 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7247ee1d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:16,373 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@670e4080{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:16,375 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:16,375 WARN [BP-729187193-172.17.0.3-1733160374064 heartbeating to localhost/127.0.0.1:35133 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:16,375 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:16,375 WARN [BP-729187193-172.17.0.3-1733160374064 heartbeating to localhost/127.0.0.1:35133 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-729187193-172.17.0.3-1733160374064 (Datanode Uuid b4a9e621-0374-407c-b999-17e2ccfe517a) service to localhost/127.0.0.1:35133 2024-12-02T17:26:16,375 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/data/data1/current/BP-729187193-172.17.0.3-1733160374064 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:16,376 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/cluster_19d236f7-cc16-b13a-262d-e9f9b9ff3850/data/data2/current/BP-729187193-172.17.0.3-1733160374064 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:16,376 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:16,382 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1dc8c743{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:26:16,382 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@6c4717a0{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:16,382 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:16,383 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1ae773f8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:16,383 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6a69944b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:16,391 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.log.dir so I do NOT create it in target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0dac6542-9ff6-7464-9d68-a58186a7f82f/hadoop.tmp.dir so I do NOT create it in target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786, deleteOnExit=true 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/test.cache.data in system properties and HBase conf 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T17:26:16,413 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir in system properties and HBase conf 2024-12-02T17:26:16,414 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T17:26:16,414 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T17:26:16,414 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T17:26:16,414 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T17:26:16,414 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:26:16,414 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:26:16,415 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T17:26:16,415 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:26:16,415 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T17:26:16,415 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T17:26:16,415 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:26:16,415 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:26:16,415 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T17:26:16,416 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/nfs.dump.dir in system properties and HBase conf 2024-12-02T17:26:16,416 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/java.io.tmpdir in system properties and HBase conf 2024-12-02T17:26:16,416 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:26:16,416 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T17:26:16,416 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T17:26:16,430 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:26:16,499 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:16,507 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:16,512 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:16,512 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:16,512 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:26:16,513 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:16,513 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@17c48ca{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:16,514 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@35c95cb4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:16,632 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5519c514{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/java.io.tmpdir/jetty-localhost-42563-hadoop-hdfs-3_4_1-tests_jar-_-any-8122933995611780231/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:26:16,633 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@47f2ada2{HTTP/1.1, (http/1.1)}{localhost:42563} 2024-12-02T17:26:16,633 INFO [Time-limited test {}] server.Server(415): Started @104305ms 2024-12-02T17:26:16,648 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:26:16,727 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:16,733 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:16,737 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:16,737 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:16,737 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:26:16,737 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7c3d2a60{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:16,738 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5917cb43{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:16,854 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1aa07d80{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/java.io.tmpdir/jetty-localhost-36433-hadoop-hdfs-3_4_1-tests_jar-_-any-12119490636073015887/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:16,854 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@7c814f59{HTTP/1.1, (http/1.1)}{localhost:36433} 2024-12-02T17:26:16,855 INFO [Time-limited test {}] server.Server(415): Started @104527ms 2024-12-02T17:26:16,856 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:16,892 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:16,895 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:16,896 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:16,896 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:16,896 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:26:16,897 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@198c3788{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:16,897 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6471b09b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:16,967 WARN [Thread-656 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data1/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:16,967 WARN [Thread-657 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data2/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:16,993 WARN [Thread-635 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:16,996 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x88a50b6633713fb0 with lease ID 0x32d3e7d6b46bf24a: Processing first storage report for DS-0338ef0e-e232-4469-8f02-5fdc748ca069 from datanode DatanodeRegistration(127.0.0.1:39625, datanodeUuid=052d4a25-70c4-4c8b-9084-08f18a3781d4, infoPort=46385, infoSecurePort=0, ipcPort=35081, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:16,996 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x88a50b6633713fb0 with lease ID 0x32d3e7d6b46bf24a: from storage DS-0338ef0e-e232-4469-8f02-5fdc748ca069 node DatanodeRegistration(127.0.0.1:39625, datanodeUuid=052d4a25-70c4-4c8b-9084-08f18a3781d4, infoPort=46385, infoSecurePort=0, ipcPort=35081, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:16,997 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x88a50b6633713fb0 with lease ID 0x32d3e7d6b46bf24a: Processing first storage report for DS-fde91fad-41c8-4640-979c-8b126363ca19 from datanode DatanodeRegistration(127.0.0.1:39625, datanodeUuid=052d4a25-70c4-4c8b-9084-08f18a3781d4, infoPort=46385, infoSecurePort=0, ipcPort=35081, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:16,997 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x88a50b6633713fb0 with lease ID 0x32d3e7d6b46bf24a: from storage DS-fde91fad-41c8-4640-979c-8b126363ca19 node DatanodeRegistration(127.0.0.1:39625, datanodeUuid=052d4a25-70c4-4c8b-9084-08f18a3781d4, infoPort=46385, infoSecurePort=0, ipcPort=35081, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:17,031 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@55c8142a{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/java.io.tmpdir/jetty-localhost-34831-hadoop-hdfs-3_4_1-tests_jar-_-any-12822042593203476183/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:17,032 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@65f2c48f{HTTP/1.1, (http/1.1)}{localhost:34831} 2024-12-02T17:26:17,032 INFO [Time-limited test {}] server.Server(415): Started @104704ms 2024-12-02T17:26:17,033 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:17,105 INFO [regionserver/3b3375d1a925:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:26:17,127 WARN [Thread-682 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data3/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:17,127 WARN [Thread-683 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data4/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:17,144 WARN [Thread-671 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:17,147 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x441f60cf5b4dc319 with lease ID 0x32d3e7d6b46bf24b: Processing first storage report for DS-ee527e74-ad21-4330-b74a-ac6397764fd4 from datanode DatanodeRegistration(127.0.0.1:36611, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=43019, infoSecurePort=0, ipcPort=44211, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:17,147 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x441f60cf5b4dc319 with lease ID 0x32d3e7d6b46bf24b: from storage DS-ee527e74-ad21-4330-b74a-ac6397764fd4 node DatanodeRegistration(127.0.0.1:36611, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=43019, infoSecurePort=0, ipcPort=44211, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:17,147 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x441f60cf5b4dc319 with lease ID 0x32d3e7d6b46bf24b: Processing first storage report for DS-3c29e8b5-6830-4a5e-9080-c2081db2d1db from datanode DatanodeRegistration(127.0.0.1:36611, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=43019, infoSecurePort=0, ipcPort=44211, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:17,147 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x441f60cf5b4dc319 with lease ID 0x32d3e7d6b46bf24b: from storage DS-3c29e8b5-6830-4a5e-9080-c2081db2d1db node DatanodeRegistration(127.0.0.1:36611, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=43019, infoSecurePort=0, ipcPort=44211, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:17,168 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535 2024-12-02T17:26:17,171 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/zookeeper_0, clientPort=58139, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T17:26:17,171 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=58139 2024-12-02T17:26:17,172 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:17,173 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:17,182 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:26:17,183 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:26:17,184 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca with version=8 2024-12-02T17:26:17,184 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase-staging 2024-12-02T17:26:17,186 INFO [Time-limited test {}] client.ConnectionUtils(128): master/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:26:17,186 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:17,186 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:17,186 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:26:17,186 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:17,186 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:26:17,187 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T17:26:17,187 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:26:17,187 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:45913 2024-12-02T17:26:17,189 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:45913 connecting to ZooKeeper ensemble=127.0.0.1:58139 2024-12-02T17:26:17,194 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:459130x0, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:26:17,194 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:45913-0x1009c06a5f00000 connected 2024-12-02T17:26:17,213 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:17,214 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:17,217 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:17,217 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca, hbase.cluster.distributed=false 2024-12-02T17:26:17,219 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:26:17,220 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45913 2024-12-02T17:26:17,222 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45913 2024-12-02T17:26:17,222 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45913 2024-12-02T17:26:17,222 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45913 2024-12-02T17:26:17,222 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45913 2024-12-02T17:26:17,246 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:26:17,247 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:17,247 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:17,247 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:26:17,247 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:17,247 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:26:17,247 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T17:26:17,247 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:26:17,248 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:37967 2024-12-02T17:26:17,250 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:37967 connecting to ZooKeeper ensemble=127.0.0.1:58139 2024-12-02T17:26:17,250 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:17,253 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:17,257 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:379670x0, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:26:17,258 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37967-0x1009c06a5f00001 connected 2024-12-02T17:26:17,258 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:17,258 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T17:26:17,259 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T17:26:17,259 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T17:26:17,261 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:26:17,264 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37967 2024-12-02T17:26:17,265 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37967 2024-12-02T17:26:17,266 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37967 2024-12-02T17:26:17,268 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37967 2024-12-02T17:26:17,268 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37967 2024-12-02T17:26:17,287 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;3b3375d1a925:45913 2024-12-02T17:26:17,288 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/3b3375d1a925,45913,1733160377186 2024-12-02T17:26:17,289 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:26:17,289 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:26:17,290 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/3b3375d1a925,45913,1733160377186 2024-12-02T17:26:17,291 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,292 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T17:26:17,293 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/3b3375d1a925,45913,1733160377186 from backup master directory 2024-12-02T17:26:17,293 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T17:26:17,293 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,294 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/3b3375d1a925,45913,1733160377186 2024-12-02T17:26:17,294 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:26:17,294 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:26:17,294 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(347): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Unable to get data of znode /hbase/backup-masters/3b3375d1a925,45913,1733160377186 because node does not exist (not an error) 2024-12-02T17:26:17,294 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:26:17,294 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=3b3375d1a925,45913,1733160377186 2024-12-02T17:26:17,303 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/hbase.id] with ID: 2391cd47-c917-41bf-989c-fca9e4c70bbd 2024-12-02T17:26:17,304 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/.tmp/hbase.id 2024-12-02T17:26:17,314 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:26:17,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:26:17,315 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/.tmp/hbase.id]:[hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/hbase.id] 2024-12-02T17:26:17,330 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:17,331 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T17:26:17,332 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 1ms. 2024-12-02T17:26:17,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:26:17,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:26:17,343 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:26:17,344 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T17:26:17,345 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:26:17,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:26:17,355 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:26:17,356 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store 2024-12-02T17:26:17,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:26:17,366 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:26:17,367 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:17,367 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:26:17,367 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:17,367 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:17,368 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:26:17,368 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:17,368 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:17,368 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160377367Disabling compacts and flushes for region at 1733160377367Disabling writes for close at 1733160377368 (+1 ms)Writing region close event to WAL at 1733160377368Closed at 1733160377368 2024-12-02T17:26:17,369 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/.initializing 2024-12-02T17:26:17,369 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186 2024-12-02T17:26:17,372 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C45913%2C1733160377186, suffix=, logDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186, archiveDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/oldWALs, maxLogs=10 2024-12-02T17:26:17,373 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C45913%2C1733160377186.1733160377372 2024-12-02T17:26:17,378 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 2024-12-02T17:26:17,383 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46385:46385),(127.0.0.1/127.0.0.1:43019:43019)] 2024-12-02T17:26:17,387 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:26:17,387 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:17,387 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,387 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,393 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,394 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T17:26:17,395 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,395 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,395 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,397 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T17:26:17,397 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,398 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:26:17,398 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,400 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T17:26:17,400 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,400 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:26:17,400 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,402 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T17:26:17,402 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,403 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:26:17,403 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,404 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,404 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,406 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,406 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,407 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T17:26:17,408 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:26:17,411 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:26:17,412 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=784028, jitterRate=-0.0030579566955566406}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T17:26:17,413 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733160377388Initializing all the Stores at 1733160377389 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160377389Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160377392 (+3 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160377392Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160377392Cleaning up temporary data from old regions at 1733160377406 (+14 ms)Region opened successfully at 1733160377413 (+7 ms) 2024-12-02T17:26:17,414 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T17:26:17,418 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@57c4fe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:26:17,419 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T17:26:17,419 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T17:26:17,419 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T17:26:17,419 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T17:26:17,420 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T17:26:17,420 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-02T17:26:17,420 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T17:26:17,423 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T17:26:17,424 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T17:26:17,425 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T17:26:17,425 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T17:26:17,426 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T17:26:17,427 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T17:26:17,427 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T17:26:17,429 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T17:26:17,430 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T17:26:17,431 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T17:26:17,433 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T17:26:17,434 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T17:26:17,436 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T17:26:17,437 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:17,438 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,438 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:17,438 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,438 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=3b3375d1a925,45913,1733160377186, sessionid=0x1009c06a5f00000, setting cluster-up flag (Was=false) 2024-12-02T17:26:17,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,446 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T17:26:17,447 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,45913,1733160377186 2024-12-02T17:26:17,450 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,450 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,456 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T17:26:17,457 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,45913,1733160377186 2024-12-02T17:26:17,458 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T17:26:17,461 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T17:26:17,461 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T17:26:17,461 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T17:26:17,461 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 3b3375d1a925,45913,1733160377186 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T17:26:17,463 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:26:17,463 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:26:17,463 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:26:17,463 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:26:17,463 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/3b3375d1a925:0, corePoolSize=10, maxPoolSize=10 2024-12-02T17:26:17,463 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,463 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:26:17,464 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,469 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:26:17,470 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T17:26:17,471 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,471 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T17:26:17,473 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733160407473 2024-12-02T17:26:17,473 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T17:26:17,473 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T17:26:17,474 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T17:26:17,474 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T17:26:17,474 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(746): ClusterId : 2391cd47-c917-41bf-989c-fca9e4c70bbd 2024-12-02T17:26:17,474 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T17:26:17,474 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T17:26:17,474 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T17:26:17,476 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,476 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T17:26:17,476 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T17:26:17,479 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T17:26:17,479 DEBUG [RS:0;3b3375d1a925:37967 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1bb792b8, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:26:17,484 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T17:26:17,484 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T17:26:17,484 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T17:26:17,486 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T17:26:17,486 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T17:26:17,489 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160377486,5,FailOnTimeoutGroup] 2024-12-02T17:26:17,489 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160377489,5,FailOnTimeoutGroup] 2024-12-02T17:26:17,489 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,489 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T17:26:17,489 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,489 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:26:17,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:26:17,492 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T17:26:17,493 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca 2024-12-02T17:26:17,500 DEBUG [RS:0;3b3375d1a925:37967 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;3b3375d1a925:37967 2024-12-02T17:26:17,500 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T17:26:17,500 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T17:26:17,500 DEBUG [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T17:26:17,501 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(2659): reportForDuty to master=3b3375d1a925,45913,1733160377186 with port=37967, startcode=1733160377246 2024-12-02T17:26:17,501 DEBUG [RS:0;3b3375d1a925:37967 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T17:26:17,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:26:17,504 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:26:17,504 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:17,506 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45147, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T17:26:17,506 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=45913 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,506 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=45913 {}] master.ServerManager(517): Registering regionserver=3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,509 DEBUG [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca 2024-12-02T17:26:17,509 DEBUG [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41131 2024-12-02T17:26:17,509 DEBUG [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T17:26:17,511 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:26:17,511 DEBUG [RS:0;3b3375d1a925:37967 {}] zookeeper.ZKUtil(111): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,511 WARN [RS:0;3b3375d1a925:37967 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:26:17,511 INFO [RS:0;3b3375d1a925:37967 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:26:17,512 DEBUG [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,513 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:26:17,513 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [3b3375d1a925,37967,1733160377246] 2024-12-02T17:26:17,516 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T17:26:17,517 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:26:17,518 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,518 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,518 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:26:17,518 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T17:26:17,519 INFO [RS:0;3b3375d1a925:37967 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:26:17,519 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,520 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:26:17,520 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,521 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,521 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:26:17,522 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:26:17,522 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,523 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,523 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:26:17,524 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T17:26:17,524 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:26:17,524 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,525 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,525 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:26:17,526 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740 2024-12-02T17:26:17,526 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740 2024-12-02T17:26:17,527 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:26:17,528 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:26:17,528 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:26:17,529 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T17:26:17,529 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,530 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:26:17,530 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:17,531 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:26:17,531 DEBUG [RS:0;3b3375d1a925:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:26:17,532 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,532 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,533 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,533 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,533 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,533 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37967,1733160377246-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:26:17,533 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:26:17,534 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=832326, jitterRate=0.05835728347301483}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:26:17,535 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733160377504Initializing all the Stores at 1733160377506 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160377506Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160377512 (+6 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160377512Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160377512Cleaning up temporary data from old regions at 1733160377528 (+16 ms)Region opened successfully at 1733160377534 (+6 ms) 2024-12-02T17:26:17,535 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:26:17,535 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:26:17,535 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:26:17,535 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:26:17,535 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:26:17,537 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:26:17,537 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160377535Disabling compacts and flushes for region at 1733160377535Disabling writes for close at 1733160377535Writing region close event to WAL at 1733160377537 (+2 ms)Closed at 1733160377537 2024-12-02T17:26:17,539 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:26:17,539 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T17:26:17,539 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T17:26:17,541 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:26:17,542 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T17:26:17,553 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T17:26:17,553 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37967,1733160377246-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,553 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,553 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.Replication(171): 3b3375d1a925,37967,1733160377246 started 2024-12-02T17:26:17,568 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,569 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1482): Serving as 3b3375d1a925,37967,1733160377246, RpcServer on 3b3375d1a925/172.17.0.3:37967, sessionid=0x1009c06a5f00001 2024-12-02T17:26:17,569 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T17:26:17,569 DEBUG [RS:0;3b3375d1a925:37967 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,569 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,37967,1733160377246' 2024-12-02T17:26:17,569 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T17:26:17,569 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T17:26:17,570 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T17:26:17,570 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T17:26:17,570 DEBUG [RS:0;3b3375d1a925:37967 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,570 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,37967,1733160377246' 2024-12-02T17:26:17,570 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T17:26:17,570 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T17:26:17,571 DEBUG [RS:0;3b3375d1a925:37967 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T17:26:17,571 INFO [RS:0;3b3375d1a925:37967 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T17:26:17,571 INFO [RS:0;3b3375d1a925:37967 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T17:26:17,673 INFO [RS:0;3b3375d1a925:37967 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C37967%2C1733160377246, suffix=, logDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246, archiveDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs, maxLogs=32 2024-12-02T17:26:17,674 INFO [RS:0;3b3375d1a925:37967 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37967%2C1733160377246.1733160377674 2024-12-02T17:26:17,681 INFO [RS:0;3b3375d1a925:37967 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 2024-12-02T17:26:17,689 DEBUG [RS:0;3b3375d1a925:37967 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43019:43019),(127.0.0.1/127.0.0.1:46385:46385)] 2024-12-02T17:26:17,693 DEBUG [3b3375d1a925:45913 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T17:26:17,693 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,695 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,37967,1733160377246, state=OPENING 2024-12-02T17:26:17,696 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T17:26:17,698 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,698 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:17,699 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:26:17,699 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:26:17,699 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:26:17,699 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,37967,1733160377246}] 2024-12-02T17:26:17,852 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T17:26:17,854 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41073, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T17:26:17,858 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T17:26:17,858 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:26:17,860 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C37967%2C1733160377246.meta, suffix=.meta, logDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246, archiveDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs, maxLogs=32 2024-12-02T17:26:17,861 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta 2024-12-02T17:26:17,867 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta 2024-12-02T17:26:17,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:26:17,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:26:17,868 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling 2024-12-02T17:26:17,868 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46385:46385),(127.0.0.1/127.0.0.1:43019:43019)] 2024-12-02T17:26:17,869 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:26:17,870 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T17:26:17,870 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T17:26:17,870 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T17:26:17,870 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T17:26:17,870 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:17,870 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T17:26:17,870 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T17:26:17,872 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:26:17,873 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:26:17,873 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,874 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,874 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:26:17,875 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:26:17,875 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,875 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,876 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:26:17,877 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:26:17,877 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,877 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,877 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:26:17,878 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:26:17,878 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:17,879 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:26:17,879 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:26:17,880 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740 2024-12-02T17:26:17,881 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740 2024-12-02T17:26:17,882 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:26:17,883 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:26:17,883 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:26:17,885 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:26:17,886 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=802990, jitterRate=0.021054625511169434}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:26:17,886 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T17:26:17,887 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733160377870Writing region info on filesystem at 1733160377870Initializing all the Stores at 1733160377871 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160377871Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160377872 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160377872Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160377872Cleaning up temporary data from old regions at 1733160377883 (+11 ms)Running coprocessor post-open hooks at 1733160377886 (+3 ms)Region opened successfully at 1733160377887 (+1 ms) 2024-12-02T17:26:17,889 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733160377852 2024-12-02T17:26:17,892 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T17:26:17,892 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T17:26:17,893 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,894 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,37967,1733160377246, state=OPEN 2024-12-02T17:26:17,899 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:26:17,899 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:26:17,899 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=3b3375d1a925,37967,1733160377246 2024-12-02T17:26:17,899 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:26:17,900 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:26:17,903 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T17:26:17,903 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,37967,1733160377246 in 200 msec 2024-12-02T17:26:17,907 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T17:26:17,907 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 364 msec 2024-12-02T17:26:17,908 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:26:17,908 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T17:26:17,909 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:26:17,910 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,37967,1733160377246, seqNum=-1] 2024-12-02T17:26:17,910 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:26:17,911 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51229, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:26:17,917 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 456 msec 2024-12-02T17:26:17,918 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733160377918, completionTime=-1 2024-12-02T17:26:17,918 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T17:26:17,918 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T17:26:17,919 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=1 2024-12-02T17:26:17,919 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733160437919 2024-12-02T17:26:17,920 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733160497919 2024-12-02T17:26:17,920 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 1 msec 2024-12-02T17:26:17,920 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,45913,1733160377186-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,920 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,45913,1733160377186-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,920 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,45913,1733160377186-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,920 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-3b3375d1a925:45913, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,920 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,921 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,922 DEBUG [master/3b3375d1a925:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T17:26:17,924 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.630sec 2024-12-02T17:26:17,924 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T17:26:17,924 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T17:26:17,924 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T17:26:17,924 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T17:26:17,924 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T17:26:17,924 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,45913,1733160377186-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:26:17,924 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,45913,1733160377186-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T17:26:17,927 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T17:26:17,927 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T17:26:17,927 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,45913,1733160377186-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:17,974 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@10c7ae56, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:26:17,974 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 3b3375d1a925,45913,-1 for getting cluster id 2024-12-02T17:26:17,974 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T17:26:17,976 DEBUG [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '2391cd47-c917-41bf-989c-fca9e4c70bbd' 2024-12-02T17:26:17,976 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T17:26:17,976 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "2391cd47-c917-41bf-989c-fca9e4c70bbd" 2024-12-02T17:26:17,977 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5d5adab1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:26:17,977 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [3b3375d1a925,45913,-1] 2024-12-02T17:26:17,977 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T17:26:17,977 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:17,979 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54776, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T17:26:17,980 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@272348fe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:26:17,981 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:26:17,982 DEBUG [RPCClient-NioEventLoopGroup-4-9 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,37967,1733160377246, seqNum=-1] 2024-12-02T17:26:17,982 DEBUG [RPCClient-NioEventLoopGroup-4-9 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:26:17,984 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:33848, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:26:17,986 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=3b3375d1a925,45913,1733160377186 2024-12-02T17:26:17,986 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:17,989 INFO [Time-limited test {}] master.MasterRpcServices(567): Client=null/null set balanceSwitch=false 2024-12-02T17:26:18,006 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:26:18,006 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:18,006 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:18,006 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:26:18,006 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:26:18,006 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:26:18,006 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T17:26:18,006 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:26:18,007 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:37629 2024-12-02T17:26:18,008 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:37629 connecting to ZooKeeper ensemble=127.0.0.1:58139 2024-12-02T17:26:18,009 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:18,011 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:18,016 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:376290x0, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:26:18,017 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37629-0x1009c06a5f00002 connected 2024-12-02T17:26:18,017 DEBUG [Time-limited test {}] zookeeper.ZKUtil(111): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on existing znode=/hbase/running 2024-12-02T17:26:18,017 DEBUG [pool-381-thread-1 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: INIT 2024-12-02T17:26:18,017 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T17:26:18,018 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T17:26:18,018 DEBUG [Time-limited test {}] zookeeper.ZKUtil(111): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T17:26:18,020 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:26:18,023 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37629 2024-12-02T17:26:18,024 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37629 2024-12-02T17:26:18,026 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37629 2024-12-02T17:26:18,026 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37629 2024-12-02T17:26:18,026 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37629 2024-12-02T17:26:18,028 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(746): ClusterId : 2391cd47-c917-41bf-989c-fca9e4c70bbd 2024-12-02T17:26:18,028 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T17:26:18,032 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T17:26:18,033 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T17:26:18,035 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T17:26:18,035 DEBUG [RS:1;3b3375d1a925:37629 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@c8abfd6, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:26:18,048 DEBUG [RS:1;3b3375d1a925:37629 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;3b3375d1a925:37629 2024-12-02T17:26:18,048 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T17:26:18,048 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T17:26:18,048 DEBUG [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T17:26:18,049 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(2659): reportForDuty to master=3b3375d1a925,45913,1733160377186 with port=37629, startcode=1733160378005 2024-12-02T17:26:18,049 DEBUG [RS:1;3b3375d1a925:37629 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T17:26:18,051 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42609, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T17:26:18,051 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=45913 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 3b3375d1a925,37629,1733160378005 2024-12-02T17:26:18,051 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=45913 {}] master.ServerManager(517): Registering regionserver=3b3375d1a925,37629,1733160378005 2024-12-02T17:26:18,053 DEBUG [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca 2024-12-02T17:26:18,053 DEBUG [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41131 2024-12-02T17:26:18,053 DEBUG [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T17:26:18,055 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:26:18,055 DEBUG [RS:1;3b3375d1a925:37629 {}] zookeeper.ZKUtil(111): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/3b3375d1a925,37629,1733160378005 2024-12-02T17:26:18,055 WARN [RS:1;3b3375d1a925:37629 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:26:18,055 INFO [RS:1;3b3375d1a925:37629 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:26:18,055 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [3b3375d1a925,37629,1733160378005] 2024-12-02T17:26:18,055 DEBUG [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005 2024-12-02T17:26:18,059 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T17:26:18,062 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T17:26:18,062 INFO [RS:1;3b3375d1a925:37629 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:26:18,062 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,062 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T17:26:18,063 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T17:26:18,063 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,063 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,063 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,063 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,063 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,063 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:26:18,064 DEBUG [RS:1;3b3375d1a925:37629 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:26:18,064 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,064 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,064 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,065 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,065 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,065 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37629,1733160378005-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:26:18,080 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T17:26:18,080 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,37629,1733160378005-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,080 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,080 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.Replication(171): 3b3375d1a925,37629,1733160378005 started 2024-12-02T17:26:18,123 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:26:18,123 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(1482): Serving as 3b3375d1a925,37629,1733160378005, RpcServer on 3b3375d1a925/172.17.0.3:37629, sessionid=0x1009c06a5f00002 2024-12-02T17:26:18,123 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T17:26:18,123 DEBUG [RS:1;3b3375d1a925:37629 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 3b3375d1a925,37629,1733160378005 2024-12-02T17:26:18,123 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,37629,1733160378005' 2024-12-02T17:26:18,123 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T17:26:18,123 INFO [Time-limited test {}] hbase.HBaseTestingUtil(2882): Started new server=Thread[RS:1;3b3375d1a925:37629,5,FailOnTimeoutGroup] 2024-12-02T17:26:18,124 INFO [Time-limited test {}] wal.TestLogRolling(207): Replication=2 2024-12-02T17:26:18,124 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-02T17:26:18,124 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T17:26:18,125 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T17:26:18,125 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T17:26:18,125 DEBUG [RS:1;3b3375d1a925:37629 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 3b3375d1a925,37629,1733160378005 2024-12-02T17:26:18,125 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,37629,1733160378005' 2024-12-02T17:26:18,125 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T17:26:18,125 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T17:26:18,125 DEBUG [RPCClient-NioEventLoopGroup-4-9 {}] client.AsyncConnectionImpl(321): The fetched master address is 3b3375d1a925,45913,1733160377186 2024-12-02T17:26:18,126 DEBUG [RPCClient-NioEventLoopGroup-4-9 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@7bc25f6f 2024-12-02T17:26:18,126 DEBUG [RS:1;3b3375d1a925:37629 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T17:26:18,126 DEBUG [RPCClient-NioEventLoopGroup-4-9 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T17:26:18,126 INFO [RS:1;3b3375d1a925:37629 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T17:26:18,126 INFO [RS:1;3b3375d1a925:37629 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T17:26:18,128 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54788, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T17:26:18,128 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45913 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T17:26:18,129 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45913 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T17:26:18,129 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45913 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestLogRolling-testLogRollOnDatanodeDeath', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:26:18,130 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45913 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath 2024-12-02T17:26:18,132 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T17:26:18,132 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:18,132 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45913 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRollOnDatanodeDeath" procId is: 4 2024-12-02T17:26:18,133 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45913 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:26:18,133 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T17:26:18,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741835_1011 (size=393) 2024-12-02T17:26:18,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741835_1011 (size=393) 2024-12-02T17:26:18,143 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 90038ff24035c0472c575f3364061f59, NAME => 'TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRollOnDatanodeDeath', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca 2024-12-02T17:26:18,149 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36611 is added to blk_1073741836_1012 (size=76) 2024-12-02T17:26:18,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39625 is added to blk_1073741836_1012 (size=76) 2024-12-02T17:26:18,150 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(898): Instantiated TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:18,150 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1722): Closing 90038ff24035c0472c575f3364061f59, disabling compactions & flushes 2024-12-02T17:26:18,150 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1755): Closing region TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:18,151 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:18,151 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. after waiting 0 ms 2024-12-02T17:26:18,151 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:18,151 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1973): Closed TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:18,151 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1676): Region close journal for 90038ff24035c0472c575f3364061f59: Waiting for close lock at 1733160378150Disabling compacts and flushes for region at 1733160378150Disabling writes for close at 1733160378151 (+1 ms)Writing region close event to WAL at 1733160378151Closed at 1733160378151 2024-12-02T17:26:18,152 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T17:26:18,153 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.","families":{"info":[{"qualifier":"regioninfo","vlen":75,"tag":[],"timestamp":"1733160378152"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733160378152"}]},"ts":"1733160378152"} 2024-12-02T17:26:18,155 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-02T17:26:18,157 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T17:26:18,157 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnDatanodeDeath","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160378157"}]},"ts":"1733160378157"} 2024-12-02T17:26:18,159 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testLogRollOnDatanodeDeath, state=ENABLING in hbase:meta 2024-12-02T17:26:18,159 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=90038ff24035c0472c575f3364061f59, ASSIGN}] 2024-12-02T17:26:18,161 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=90038ff24035c0472c575f3364061f59, ASSIGN 2024-12-02T17:26:18,162 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=90038ff24035c0472c575f3364061f59, ASSIGN; state=OFFLINE, location=3b3375d1a925,37967,1733160377246; forceNewPlan=false, retain=false 2024-12-02T17:26:18,228 INFO [RS:1;3b3375d1a925:37629 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C37629%2C1733160378005, suffix=, logDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005, archiveDir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs, maxLogs=32 2024-12-02T17:26:18,229 INFO [RS:1;3b3375d1a925:37629 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37629%2C1733160378005.1733160378229 2024-12-02T17:26:18,237 INFO [RS:1;3b3375d1a925:37629 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005/3b3375d1a925%2C37629%2C1733160378005.1733160378229 2024-12-02T17:26:18,238 DEBUG [RS:1;3b3375d1a925:37629 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43019:43019),(127.0.0.1/127.0.0.1:46385:46385)] 2024-12-02T17:26:18,313 INFO [3b3375d1a925:45913 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-02T17:26:18,313 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=90038ff24035c0472c575f3364061f59, regionState=OPENING, regionLocation=3b3375d1a925,37967,1733160377246 2024-12-02T17:26:18,316 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=90038ff24035c0472c575f3364061f59, ASSIGN because future has completed 2024-12-02T17:26:18,316 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 90038ff24035c0472c575f3364061f59, server=3b3375d1a925,37967,1733160377246}] 2024-12-02T17:26:18,473 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:18,473 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 90038ff24035c0472c575f3364061f59, NAME => 'TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:26:18,474 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRollOnDatanodeDeath 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,474 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:26:18,474 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,474 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,476 INFO [StoreOpener-90038ff24035c0472c575f3364061f59-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,477 INFO [StoreOpener-90038ff24035c0472c575f3364061f59-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 90038ff24035c0472c575f3364061f59 columnFamilyName info 2024-12-02T17:26:18,477 DEBUG [StoreOpener-90038ff24035c0472c575f3364061f59-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:26:18,478 INFO [StoreOpener-90038ff24035c0472c575f3364061f59-1 {}] regionserver.HStore(327): Store=90038ff24035c0472c575f3364061f59/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:26:18,478 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,478 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,479 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,479 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,479 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,481 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,483 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:26:18,483 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 90038ff24035c0472c575f3364061f59; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=780787, jitterRate=-0.0071784257888793945}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T17:26:18,483 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:18,484 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 90038ff24035c0472c575f3364061f59: Running coprocessor pre-open hook at 1733160378474Writing region info on filesystem at 1733160378474Initializing all the Stores at 1733160378475 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160378475Cleaning up temporary data from old regions at 1733160378479 (+4 ms)Running coprocessor post-open hooks at 1733160378483 (+4 ms)Region opened successfully at 1733160378484 (+1 ms) 2024-12-02T17:26:18,484 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:18,485 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59., pid=6, masterSystemTime=1733160378469 2024-12-02T17:26:18,488 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:18,488 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:18,488 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:18,489 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=90038ff24035c0472c575f3364061f59, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,37967,1733160377246 2024-12-02T17:26:18,492 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 90038ff24035c0472c575f3364061f59, server=3b3375d1a925,37967,1733160377246 because future has completed 2024-12-02T17:26:18,497 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T17:26:18,497 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 90038ff24035c0472c575f3364061f59, server=3b3375d1a925,37967,1733160377246 in 177 msec 2024-12-02T17:26:18,500 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T17:26:18,500 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=90038ff24035c0472c575f3364061f59, ASSIGN in 338 msec 2024-12-02T17:26:18,501 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T17:26:18,502 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnDatanodeDeath","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160378501"}]},"ts":"1733160378501"} 2024-12-02T17:26:18,504 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testLogRollOnDatanodeDeath, state=ENABLED in hbase:meta 2024-12-02T17:26:18,505 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T17:26:18,508 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath in 376 msec 2024-12-02T17:26:19,003 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:26:19,005 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:19,028 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:19,029 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:19,030 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:23,976 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:26:23,978 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:24,006 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:24,007 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:24,008 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:24,018 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T17:26:24,019 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRollOnDatanodeDeath' 2024-12-02T17:26:27,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:26:27,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-02T17:26:27,868 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnDatanodeDeath 2024-12-02T17:26:27,868 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnDatanodeDeath Metrics about Tables on a single HBase RegionServer 2024-12-02T17:26:27,869 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:26:27,869 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-02T17:26:28,169 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45913 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:26:28,169 INFO [RPCClient-NioEventLoopGroup-4-11 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestLogRolling-testLogRollOnDatanodeDeath completed 2024-12-02T17:26:28,170 DEBUG [Time-limited test {}] hbase.ClientMetaTableAccessor(255): Scanning META starting at row=TestLogRolling-testLogRollOnDatanodeDeath,, stopping at row=TestLogRolling-testLogRollOnDatanodeDeath ,, for max=2147483647 with caching=100 2024-12-02T17:26:28,173 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2234): Found 1 regions for table TestLogRolling-testLogRollOnDatanodeDeath 2024-12-02T17:26:28,173 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2240): firstRegionName=TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:28,187 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:28,190 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:28,192 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:28,192 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:28,192 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:26:28,192 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@219c70cc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:28,193 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@75255721{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:28,306 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@22d0350b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/java.io.tmpdir/jetty-localhost-40885-hadoop-hdfs-3_4_1-tests_jar-_-any-6463622674815449767/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:28,307 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3cc6081e{HTTP/1.1, (http/1.1)}{localhost:40885} 2024-12-02T17:26:28,307 INFO [Time-limited test {}] server.Server(415): Started @115979ms 2024-12-02T17:26:28,308 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:28,338 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:28,342 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:28,344 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:28,344 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:28,345 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:26:28,345 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4fb99827{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:28,346 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7fbc343d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:28,404 WARN [Thread-828 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data5/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:28,404 WARN [Thread-829 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data6/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:28,422 WARN [Thread-808 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:28,425 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7d715fdd05aefb52 with lease ID 0x32d3e7d6b46bf24c: Processing first storage report for DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b from datanode DatanodeRegistration(127.0.0.1:45809, datanodeUuid=ac4b3033-1252-4511-8ac6-84eb9993023e, infoPort=35215, infoSecurePort=0, ipcPort=43189, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:28,425 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7d715fdd05aefb52 with lease ID 0x32d3e7d6b46bf24c: from storage DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b node DatanodeRegistration(127.0.0.1:45809, datanodeUuid=ac4b3033-1252-4511-8ac6-84eb9993023e, infoPort=35215, infoSecurePort=0, ipcPort=43189, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:28,425 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7d715fdd05aefb52 with lease ID 0x32d3e7d6b46bf24c: Processing first storage report for DS-085715fd-6072-44f3-9222-33563cce8ec9 from datanode DatanodeRegistration(127.0.0.1:45809, datanodeUuid=ac4b3033-1252-4511-8ac6-84eb9993023e, infoPort=35215, infoSecurePort=0, ipcPort=43189, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:28,425 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7d715fdd05aefb52 with lease ID 0x32d3e7d6b46bf24c: from storage DS-085715fd-6072-44f3-9222-33563cce8ec9 node DatanodeRegistration(127.0.0.1:45809, datanodeUuid=ac4b3033-1252-4511-8ac6-84eb9993023e, infoPort=35215, infoSecurePort=0, ipcPort=43189, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:28,464 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3df1987c{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/java.io.tmpdir/jetty-localhost-46381-hadoop-hdfs-3_4_1-tests_jar-_-any-12327429082202604198/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:28,464 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@71bfb2ba{HTTP/1.1, (http/1.1)}{localhost:46381} 2024-12-02T17:26:28,464 INFO [Time-limited test {}] server.Server(415): Started @116136ms 2024-12-02T17:26:28,465 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:28,500 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:28,503 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:28,507 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:28,507 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:28,507 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:26:28,508 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@41559526{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:28,508 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@495a6aea{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:28,576 WARN [Thread-863 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data7/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:28,576 WARN [Thread-864 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data8/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:28,600 WARN [Thread-843 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:28,603 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xfdfd7ad1483aade7 with lease ID 0x32d3e7d6b46bf24d: Processing first storage report for DS-176dfedc-f5b8-4b28-b862-d6e648f52d33 from datanode DatanodeRegistration(127.0.0.1:45891, datanodeUuid=22a41a77-5404-4648-bb58-afef246a5be7, infoPort=38693, infoSecurePort=0, ipcPort=43595, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:28,603 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xfdfd7ad1483aade7 with lease ID 0x32d3e7d6b46bf24d: from storage DS-176dfedc-f5b8-4b28-b862-d6e648f52d33 node DatanodeRegistration(127.0.0.1:45891, datanodeUuid=22a41a77-5404-4648-bb58-afef246a5be7, infoPort=38693, infoSecurePort=0, ipcPort=43595, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:28,603 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xfdfd7ad1483aade7 with lease ID 0x32d3e7d6b46bf24d: Processing first storage report for DS-ec202c66-c581-4b3b-a6a3-61a3f9cbba8e from datanode DatanodeRegistration(127.0.0.1:45891, datanodeUuid=22a41a77-5404-4648-bb58-afef246a5be7, infoPort=38693, infoSecurePort=0, ipcPort=43595, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:28,603 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xfdfd7ad1483aade7 with lease ID 0x32d3e7d6b46bf24d: from storage DS-ec202c66-c581-4b3b-a6a3-61a3f9cbba8e node DatanodeRegistration(127.0.0.1:45891, datanodeUuid=22a41a77-5404-4648-bb58-afef246a5be7, infoPort=38693, infoSecurePort=0, ipcPort=43595, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:28,639 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@f0760d{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/java.io.tmpdir/jetty-localhost-42183-hadoop-hdfs-3_4_1-tests_jar-_-any-16774312775637020428/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:28,639 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@21f536ea{HTTP/1.1, (http/1.1)}{localhost:42183} 2024-12-02T17:26:28,639 INFO [Time-limited test {}] server.Server(415): Started @116312ms 2024-12-02T17:26:28,641 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:28,724 WARN [Thread-889 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:28,724 WARN [Thread-890 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10/current/BP-797259812-172.17.0.3-1733160376449/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:28,745 WARN [Thread-878 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:28,748 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x6187f0b94ef78e2c with lease ID 0x32d3e7d6b46bf24e: Processing first storage report for DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70 from datanode DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:28,748 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6187f0b94ef78e2c with lease ID 0x32d3e7d6b46bf24e: from storage DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70 node DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:28,748 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x6187f0b94ef78e2c with lease ID 0x32d3e7d6b46bf24e: Processing first storage report for DS-bf0fbd51-74e9-453b-ad45-db99b305fd3f from datanode DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449) 2024-12-02T17:26:28,748 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6187f0b94ef78e2c with lease ID 0x32d3e7d6b46bf24e: from storage DS-bf0fbd51-74e9-453b-ad45-db99b305fd3f node DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:28,759 WARN [ResponseProcessor for block BP-797259812-172.17.0.3-1733160376449:blk_1073741837_1013 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-797259812-172.17.0.3-1733160376449:blk_1073741837_1013 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,759 WARN [ResponseProcessor for block BP-797259812-172.17.0.3-1733160376449:blk_1073741833_1009 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-797259812-172.17.0.3-1733160376449:blk_1073741833_1009 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,760 WARN [ResponseProcessor for block BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010 java.io.IOException: Bad response ERROR for BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010 from datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,760 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005/3b3375d1a925%2C37629%2C1733160378005.1733160378229 block BP-797259812-172.17.0.3-1733160376449:blk_1073741837_1013 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741837_1013 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:28,760 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta block BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:28,760 WARN [ResponseProcessor for block BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006 java.io.IOException: Bad response ERROR for BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006 from datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,760 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 block BP-797259812-172.17.0.3-1733160376449:blk_1073741833_1009 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741833_1009 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:28,760 WARN [PacketResponder: BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:36611] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,760 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 block BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:28,762 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:58778 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:39625:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:58778 dst: /127.0.0.1:39625 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,761 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:44372 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:36611:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:44372 dst: /127.0.0.1:36611 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,761 WARN [PacketResponder: BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:36611] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,762 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:44380 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:36611:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:44380 dst: /127.0.0.1:36611 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,762 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1600066942_22 at /127.0.0.1:44396 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741837_1013] {}] datanode.DataXceiver(331): 127.0.0.1:36611:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:44396 dst: /127.0.0.1:36611 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,763 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@55c8142a{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:28,763 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1600066942_22 at /127.0.0.1:58814 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741837_1013] {}] datanode.DataXceiver(331): 127.0.0.1:39625:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:58814 dst: /127.0.0.1:39625 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,763 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1129077784_22 at /127.0.0.1:58752 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:39625:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:58752 dst: /127.0.0.1:39625 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,761 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:58782 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:39625:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:58782 dst: /127.0.0.1:39625 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,763 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@65f2c48f{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:28,763 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1129077784_22 at /127.0.0.1:44342 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:36611:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:44342 dst: /127.0.0.1:36611 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,763 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:28,764 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6471b09b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:28,764 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@198c3788{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:28,765 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:28,765 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:28,765 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:28,765 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-797259812-172.17.0.3-1733160376449 (Datanode Uuid 9f811f10-21a0-40f9-b54a-2bca063deae0) service to localhost/127.0.0.1:41131 2024-12-02T17:26:28,766 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data3/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:28,766 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:28,767 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data4/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:28,767 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta block BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1010 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741834_1010 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,768 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 block BP-797259812-172.17.0.3-1733160376449:blk_1073741833_1009 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741833_1009 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,768 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 block BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1006 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741830_1006 java.net.SocketException: Connection reset at sun.nio.ch.SocketChannelImpl.throwConnectionReset(SocketChannelImpl.java:394) ~[?:?] at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:426) ~[?:?] at org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:118) ~[hadoop-common-3.4.1.jar:?] at java.io.FilterInputStream.read(FilterInputStream.java:82) ~[?:?] at java.io.FilterInputStream.read(FilterInputStream.java:82) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:527) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1931) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,769 ERROR [org.apache.hadoop.hdfs.server.datanode.DataXceiver@734fb4dd {}] datanode.DataXceiver(331): 127.0.0.1:39625:DataXceiver error processing unknown operation src: /127.0.0.1:46442 dst: /127.0.0.1:39625 java.io.IOException: Server closed. at org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.addPeer(DataXceiverServer.java:334) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:232) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:28,770 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005/3b3375d1a925%2C37629%2C1733160378005.1733160378229 block BP-797259812-172.17.0.3-1733160376449:blk_1073741837_1013 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741837_1013 java.net.SocketException: Connection reset at sun.nio.ch.SocketChannelImpl.throwConnectionReset(SocketChannelImpl.java:394) ~[?:?] at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:426) ~[?:?] at org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:118) ~[hadoop-common-3.4.1.jar:?] at java.io.FilterInputStream.read(FilterInputStream.java:82) ~[?:?] at java.io.FilterInputStream.read(FilterInputStream.java:82) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:527) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1931) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,773 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1aa07d80{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:28,774 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7c814f59{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:28,774 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:28,774 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5917cb43{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:28,774 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7c3d2a60{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:28,776 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:28,776 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:28,776 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:28,776 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-797259812-172.17.0.3-1733160376449 (Datanode Uuid 052d4a25-70c4-4c8b-9084-08f18a3781d4) service to localhost/127.0.0.1:41131 2024-12-02T17:26:28,776 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data1/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:28,777 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data2/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:28,777 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:28,781 DEBUG [RPCClient-NioEventLoopGroup-4-10 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestLogRolling-testLogRollOnDatanodeDeath', row='row0002', locateType=CURRENT is [region=TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59., hostname=3b3375d1a925,37967,1733160377246, seqNum=2] 2024-12-02T17:26:28,783 ERROR [FSHLog-0-hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca-prefix:3b3375d1a925,37967,1733160377246 {}] wal.AbstractFSWAL(1838): appendAndSync throws IOException. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,783 WARN [FSHLog-0-hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca-prefix:3b3375d1a925,37967,1733160377246 {}] wal.AbstractFSWAL(2174): append entry failed java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,784 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,784 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C37967%2C1733160377246:(num 1733160377674) roll requested 2024-12-02T17:26:28,784 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37967%2C1733160377246.1733160388784 2024-12-02T17:26:28,787 WARN [Thread-901 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741838_1018 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,788 WARN [Thread-901 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741838_1018 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:28,788 WARN [Thread-901 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741838_1018 2024-12-02T17:26:28,791 WARN [Thread-901 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:28,801 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:28,802 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:28,802 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:28,802 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:28,802 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:28,803 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 with entries=1, filesize=455 B; new WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160388784 2024-12-02T17:26:28,803 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,803 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:28,804 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(47): Initialize RecoverLeaseFSUtils 2024-12-02T17:26:28,804 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(59): set recoverLeaseMethod to org.apache.hadoop.fs.LeaseRecoverable.recoverLease() 2024-12-02T17:26:28,805 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 2024-12-02T17:26:28,807 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35215:35215),(127.0.0.1/127.0.0.1:38693:38693)] 2024-12-02T17:26:28,807 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 is not closed yet, will try archiving it next time 2024-12-02T17:26:28,807 WARN [IPC Server handler 0 on default port 41131 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 has not been closed. Lease recovery is in progress. RecoveryId = 1020 for block blk_1073741833_1009 2024-12-02T17:26:28,811 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 after 5ms 2024-12-02T17:26:29,346 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:30,065 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:30,808 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:30,809 INFO [Time-limited test {}] wal.TestLogRolling(261): log.getCurrentFileName(): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160388784 2024-12-02T17:26:30,810 WARN [ResponseProcessor for block BP-797259812-172.17.0.3-1733160376449:blk_1073741839_1019 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-797259812-172.17.0.3-1733160376449:blk_1073741839_1019 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:30,810 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160388784 block BP-797259812-172.17.0.3-1733160376449:blk_1073741839_1019 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741839_1019 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:30,811 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:46494 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741839_1019] {}] datanode.DataXceiver(331): 127.0.0.1:45809:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46494 dst: /127.0.0.1:45809 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:30,811 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:46950 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741839_1019] {}] datanode.DataXceiver(331): 127.0.0.1:45891:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46950 dst: /127.0.0.1:45891 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:30,815 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@22d0350b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:30,815 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3cc6081e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:30,815 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:30,815 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@75255721{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:30,816 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@219c70cc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:30,816 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:30,816 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:30,817 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-797259812-172.17.0.3-1733160376449 (Datanode Uuid ac4b3033-1252-4511-8ac6-84eb9993023e) service to localhost/127.0.0.1:41131 2024-12-02T17:26:30,817 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:30,817 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data5/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:30,817 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data6/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:30,818 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:31,346 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:32,065 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:32,808 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:32,808 WARN [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(529): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]] 2024-12-02T17:26:32,809 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C37967%2C1733160377246:(num 1733160388784) roll requested 2024-12-02T17:26:32,809 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37967%2C1733160377246.1733160392809 2024-12-02T17:26:32,812 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=1 on file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 after 4007ms 2024-12-02T17:26:32,815 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:32,815 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:32,815 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:32,815 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:32,815 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:32,815 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160388784 with entries=2, filesize=2.37 KB; new WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160392809 2024-12-02T17:26:32,816 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40409:40409),(127.0.0.1/127.0.0.1:38693:38693)] 2024-12-02T17:26:32,816 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 is not closed yet, will try archiving it next time 2024-12-02T17:26:32,816 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160388784 is not closed yet, will try archiving it next time 2024-12-02T17:26:32,817 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45891 is added to blk_1073741839_1021 (size=2431) 2024-12-02T17:26:32,824 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [10,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T17:26:33,218 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 is not closed yet, will try archiving it next time 2024-12-02T17:26:33,346 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:34,066 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:34,816 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:34,830 WARN [ResponseProcessor for block BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1022 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1022 java.io.IOException: Bad response ERROR for BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1022 from datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:34,830 WARN [DataStreamer for file /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160392809 block BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1022 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1022 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:34,830 WARN [PacketResponder: BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1022, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:45891] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,831 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39002 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1022] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39002 dst: /127.0.0.1:34399 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,831 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:46974 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1022] {}] datanode.DataXceiver(331): 127.0.0.1:45891:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46974 dst: /127.0.0.1:45891 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,833 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3df1987c{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:34,833 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@71bfb2ba{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:34,833 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:34,834 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7fbc343d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:34,834 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4fb99827{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:34,836 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:34,836 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:34,836 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-797259812-172.17.0.3-1733160376449 (Datanode Uuid 22a41a77-5404-4648-bb58-afef246a5be7) service to localhost/127.0.0.1:41131 2024-12-02T17:26:34,836 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:34,836 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data7/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:34,837 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data8/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:34,837 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:34,847 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8855): Flush requested on 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:34,847 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 90038ff24035c0472c575f3364061f59 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:26:34,865 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/d7340ed25fd247928cf6d5048618e1db is 1080, key is row0002/info:/1733160390821/Put/seqid=0 2024-12-02T17:26:34,868 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39012 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741841_1024] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741841_1024 to mirror 127.0.0.1:45809 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,868 WARN [Thread-919 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741841_1024 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:45809 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:34,868 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39012 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741841_1024] {}] datanode.BlockReceiver(316): Block 1073741841 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T17:26:34,868 WARN [Thread-919 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741841_1024 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:34,868 WARN [Thread-919 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741841_1024 2024-12-02T17:26:34,868 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39012 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741841_1024] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39012 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,869 WARN [Thread-919 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:34,870 WARN [Thread-919 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741842_1025 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:34,870 WARN [Thread-919 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741842_1025 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:34,870 WARN [Thread-919 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741842_1025 2024-12-02T17:26:34,871 WARN [Thread-919 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:34,874 WARN [Thread-919 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741843_1026 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:39625 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:34,874 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39018 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741843_1026] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741843_1026 to mirror 127.0.0.1:39625 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,874 WARN [Thread-919 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741843_1026 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:34,874 WARN [Thread-919 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741843_1026 2024-12-02T17:26:34,874 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39018 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741843_1026] {}] datanode.BlockReceiver(316): Block 1073741843 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T17:26:34,874 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39018 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741843_1026] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39018 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,874 WARN [Thread-919 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:34,876 WARN [Thread-919 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741844_1027 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:45891 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:34,876 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39028 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741844_1027] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741844_1027 to mirror 127.0.0.1:45891 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,877 WARN [Thread-919 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741844_1027 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:34,877 WARN [Thread-919 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741844_1027 2024-12-02T17:26:34,877 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39028 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741844_1027] {}] datanode.BlockReceiver(316): Block 1073741844 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T17:26:34,877 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39028 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741844_1027] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39028 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:34,877 WARN [Thread-919 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:34,878 WARN [IPC Server handler 3 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:34,878 WARN [IPC Server handler 3 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:34,878 WARN [IPC Server handler 3 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:34,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741845_1028 (size=10347) 2024-12-02T17:26:35,282 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/d7340ed25fd247928cf6d5048618e1db 2024-12-02T17:26:35,291 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/d7340ed25fd247928cf6d5048618e1db as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/d7340ed25fd247928cf6d5048618e1db 2024-12-02T17:26:35,296 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/d7340ed25fd247928cf6d5048618e1db, entries=5, sequenceid=11, filesize=10.1 K 2024-12-02T17:26:35,298 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=9.45 KB/9681 for 90038ff24035c0472c575f3364061f59 in 450ms, sequenceid=11, compaction requested=false 2024-12-02T17:26:35,298 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 90038ff24035c0472c575f3364061f59: 2024-12-02T17:26:35,347 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:35,470 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8855): Flush requested on 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:35,470 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 90038ff24035c0472c575f3364061f59 1/1 column families, dataSize=10.50 KB heapSize=11.50 KB 2024-12-02T17:26:35,474 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/234fe182545e476cb5e8e763fe9b490a is 1080, key is row0007/info:/1733160394848/Put/seqid=0 2024-12-02T17:26:35,476 WARN [Thread-927 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741846_1029 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:35,477 WARN [Thread-927 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741846_1029 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:35,477 WARN [Thread-927 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741846_1029 2024-12-02T17:26:35,477 WARN [Thread-927 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:35,478 WARN [Thread-927 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741847_1030 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:35,479 WARN [Thread-927 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741847_1030 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:35,479 WARN [Thread-927 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741847_1030 2024-12-02T17:26:35,479 WARN [Thread-927 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:35,480 WARN [Thread-927 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741848_1031 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:35,480 WARN [Thread-927 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741848_1031 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:35,480 WARN [Thread-927 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741848_1031 2024-12-02T17:26:35,481 WARN [Thread-927 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:35,483 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39060 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741849_1032] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741849_1032 to mirror 127.0.0.1:45809 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:35,483 WARN [Thread-927 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741849_1032 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:45809 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:35,483 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39060 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741849_1032] {}] datanode.BlockReceiver(316): Block 1073741849 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T17:26:35,483 WARN [Thread-927 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741849_1032 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:35,483 WARN [Thread-927 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741849_1032 2024-12-02T17:26:35,483 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:39060 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741849_1032] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39060 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:35,484 WARN [Thread-927 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:35,484 WARN [IPC Server handler 3 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:35,485 WARN [IPC Server handler 3 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:35,485 WARN [IPC Server handler 3 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:35,488 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741850_1033 (size=12506) 2024-12-02T17:26:35,889 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=10.50 KB at sequenceid=24 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/234fe182545e476cb5e8e763fe9b490a 2024-12-02T17:26:35,896 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/234fe182545e476cb5e8e763fe9b490a as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/234fe182545e476cb5e8e763fe9b490a 2024-12-02T17:26:35,901 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/234fe182545e476cb5e8e763fe9b490a, entries=7, sequenceid=24, filesize=12.2 K 2024-12-02T17:26:35,902 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~10.50 KB/10756, heapSize ~11.48 KB/11760, currentSize=2.10 KB/2150 for 90038ff24035c0472c575f3364061f59 in 432ms, sequenceid=24, compaction requested=false 2024-12-02T17:26:35,902 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 90038ff24035c0472c575f3364061f59: 2024-12-02T17:26:35,903 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=22.3 K, sizeToCheck=16.0 K 2024-12-02T17:26:35,903 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:35,903 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/234fe182545e476cb5e8e763fe9b490a because midkey is the same as first or last row 2024-12-02T17:26:36,066 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,817 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,817 WARN [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(529): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]] 2024-12-02T17:26:36,817 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C37967%2C1733160377246:(num 1733160392809) roll requested 2024-12-02T17:26:36,818 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37967%2C1733160377246.1733160396817 2024-12-02T17:26:36,820 WARN [Thread-932 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741851_1034 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,821 WARN [Thread-932 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741851_1034 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:36,821 WARN [Thread-932 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741851_1034 2024-12-02T17:26:36,821 WARN [Thread-932 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:36,822 WARN [Thread-932 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741852_1035 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,822 WARN [Thread-932 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741852_1035 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:36,822 WARN [Thread-932 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741852_1035 2024-12-02T17:26:36,823 WARN [Thread-932 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:36,824 WARN [Thread-932 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741853_1036 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,824 WARN [Thread-932 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741853_1036 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:36,824 WARN [Thread-932 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741853_1036 2024-12-02T17:26:36,825 WARN [Thread-932 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:36,826 WARN [Thread-932 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741854_1037 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,826 WARN [Thread-932 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741854_1037 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:36,826 WARN [Thread-932 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741854_1037 2024-12-02T17:26:36,826 WARN [Thread-932 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:36,827 WARN [IPC Server handler 4 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:36,827 WARN [IPC Server handler 4 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:36,827 WARN [IPC Server handler 4 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:36,829 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:36,829 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:36,829 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:36,829 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:36,830 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:36,830 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160392809 with entries=25, filesize=25.38 KB; new WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160396817 2024-12-02T17:26:36,831 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40409:40409)] 2024-12-02T17:26:36,831 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 is not closed yet, will try archiving it next time 2024-12-02T17:26:36,831 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160392809 is not closed yet, will try archiving it next time 2024-12-02T17:26:36,832 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160388784 to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs/3b3375d1a925%2C37967%2C1733160377246.1733160388784 2024-12-02T17:26:36,832 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741840_1023 (size=25992) 2024-12-02T17:26:36,832 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 is not closed yet, will try archiving it next time 2024-12-02T17:26:36,888 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8855): Flush requested on 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:36,889 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 90038ff24035c0472c575f3364061f59 1/1 column families, dataSize=7.35 KB heapSize=8.13 KB 2024-12-02T17:26:36,893 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/517367aec22d42aea6abc75b1cbbc5dd is 1079, key is tmprow/info:/1733160396887/Put/seqid=0 2024-12-02T17:26:36,894 WARN [Thread-936 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741856_1039 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,895 WARN [Thread-936 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741856_1039 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:36,895 WARN [Thread-936 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741856_1039 2024-12-02T17:26:36,895 WARN [Thread-936 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:36,896 WARN [Thread-936 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741857_1040 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,896 WARN [Thread-936 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741857_1040 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:36,896 WARN [Thread-936 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741857_1040 2024-12-02T17:26:36,897 WARN [Thread-936 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:36,898 WARN [Thread-936 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741858_1041 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,898 WARN [Thread-936 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741858_1041 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:36,898 WARN [Thread-936 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741858_1041 2024-12-02T17:26:36,898 WARN [Thread-936 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:36,899 WARN [Thread-936 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741859_1042 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:36,899 WARN [Thread-936 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741859_1042 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:36,899 WARN [Thread-936 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741859_1042 2024-12-02T17:26:36,900 WARN [Thread-936 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:36,900 WARN [IPC Server handler 0 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:36,900 WARN [IPC Server handler 0 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:36,900 WARN [IPC Server handler 0 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:36,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741860_1043 (size=6027) 2024-12-02T17:26:37,304 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.35 KB at sequenceid=34 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/517367aec22d42aea6abc75b1cbbc5dd 2024-12-02T17:26:37,311 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/517367aec22d42aea6abc75b1cbbc5dd as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/517367aec22d42aea6abc75b1cbbc5dd 2024-12-02T17:26:37,317 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/517367aec22d42aea6abc75b1cbbc5dd, entries=1, sequenceid=34, filesize=5.9 K 2024-12-02T17:26:37,318 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.35 KB/7525, heapSize ~8.11 KB/8304, currentSize=2.10 KB/2150 for 90038ff24035c0472c575f3364061f59 in 430ms, sequenceid=34, compaction requested=true 2024-12-02T17:26:37,318 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 90038ff24035c0472c575f3364061f59: 2024-12-02T17:26:37,319 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=28.2 K, sizeToCheck=16.0 K 2024-12-02T17:26:37,319 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:37,319 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/234fe182545e476cb5e8e763fe9b490a because midkey is the same as first or last row 2024-12-02T17:26:37,319 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 90038ff24035c0472c575f3364061f59:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:26:37,319 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:26:37,319 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:26:37,320 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 28880 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:26:37,320 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HStore(1541): 90038ff24035c0472c575f3364061f59/info is initiating minor compaction (all files) 2024-12-02T17:26:37,320 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of 90038ff24035c0472c575f3364061f59/info in TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:37,321 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/d7340ed25fd247928cf6d5048618e1db, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/234fe182545e476cb5e8e763fe9b490a, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/517367aec22d42aea6abc75b1cbbc5dd] into tmpdir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp, totalSize=28.2 K 2024-12-02T17:26:37,321 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.Compactor(225): Compacting d7340ed25fd247928cf6d5048618e1db, keycount=5, bloomtype=ROW, size=10.1 K, encoding=NONE, compression=NONE, seqNum=11, earliestPutTs=1733160390821 2024-12-02T17:26:37,321 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.Compactor(225): Compacting 234fe182545e476cb5e8e763fe9b490a, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=24, earliestPutTs=1733160394848 2024-12-02T17:26:37,322 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.Compactor(225): Compacting 517367aec22d42aea6abc75b1cbbc5dd, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=34, earliestPutTs=1733160396887 2024-12-02T17:26:37,337 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 90038ff24035c0472c575f3364061f59#info#compaction#21 average throughput is 6.16 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:26:37,337 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/c1ef748d5d91496b8e6e5262bb56d0b6 is 1080, key is row0002/info:/1733160390821/Put/seqid=0 2024-12-02T17:26:37,339 WARN [Thread-942 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741861_1044 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:37,339 WARN [Thread-942 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741861_1044 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:37,339 WARN [Thread-942 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741861_1044 2024-12-02T17:26:37,340 WARN [Thread-942 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:37,341 WARN [Thread-942 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741862_1045 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:37,341 WARN [Thread-942 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741862_1045 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:37,341 WARN [Thread-942 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741862_1045 2024-12-02T17:26:37,341 WARN [Thread-942 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:37,342 WARN [Thread-942 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741863_1046 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:37,343 WARN [Thread-942 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741863_1046 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:37,343 WARN [Thread-942 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741863_1046 2024-12-02T17:26:37,343 WARN [Thread-942 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:37,345 WARN [Thread-942 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741864_1047 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:45891 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:37,345 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49810 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741864_1047] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741864_1047 to mirror 127.0.0.1:45891 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:37,345 WARN [Thread-942 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741864_1047 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:37,345 WARN [Thread-942 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741864_1047 2024-12-02T17:26:37,345 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49810 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741864_1047] {}] datanode.BlockReceiver(316): Block 1073741864 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T17:26:37,345 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49810 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741864_1047] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49810 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:37,346 WARN [Thread-942 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:37,346 WARN [IPC Server handler 1 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:37,346 WARN [IPC Server handler 1 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:37,346 WARN [IPC Server handler 1 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:37,347 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:37,349 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741865_1048 (size=17994) 2024-12-02T17:26:37,758 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/c1ef748d5d91496b8e6e5262bb56d0b6 as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6 2024-12-02T17:26:37,761 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@a350071[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741850_1033 to 127.0.0.1:45891 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:37,761 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@2882550b[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741845_1028 to 127.0.0.1:45891 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:37,765 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in 90038ff24035c0472c575f3364061f59/info of 90038ff24035c0472c575f3364061f59 into c1ef748d5d91496b8e6e5262bb56d0b6(size=17.6 K), total size for store is 17.6 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:26:37,765 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for 90038ff24035c0472c575f3364061f59: 2024-12-02T17:26:37,766 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59., storeName=90038ff24035c0472c575f3364061f59/info, priority=13, startTime=1733160397319; duration=0sec 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=17.6 K, sizeToCheck=16.0 K 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6 because midkey is the same as first or last row 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=17.6 K, sizeToCheck=16.0 K 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6 because midkey is the same as first or last row 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=17.6 K, sizeToCheck=16.0 K 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6 because midkey is the same as first or last row 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:26:37,766 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 90038ff24035c0472c575f3364061f59:info 2024-12-02T17:26:38,066 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,307 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8855): Flush requested on 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:38,307 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 90038ff24035c0472c575f3364061f59 1/1 column families, dataSize=7.35 KB heapSize=8.13 KB 2024-12-02T17:26:38,312 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/b42aac12c04a45d08a8fbcb70def59e1 is 1079, key is tmprow/info:/1733160398306/Put/seqid=0 2024-12-02T17:26:38,315 WARN [Thread-951 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741866_1049 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:36611 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,315 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49830 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741866_1049] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741866_1049 to mirror 127.0.0.1:36611 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:38,315 WARN [Thread-951 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741866_1049 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:38,315 WARN [Thread-951 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741866_1049 2024-12-02T17:26:38,315 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49830 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741866_1049] {}] datanode.BlockReceiver(316): Block 1073741866 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T17:26:38,315 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49830 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741866_1049] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49830 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:38,316 WARN [Thread-951 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:38,317 WARN [Thread-951 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741867_1050 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,317 WARN [Thread-951 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741867_1050 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:38,317 WARN [Thread-951 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741867_1050 2024-12-02T17:26:38,317 WARN [Thread-951 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:38,318 WARN [Thread-951 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741868_1051 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,319 WARN [Thread-951 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741868_1051 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:38,319 WARN [Thread-951 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741868_1051 2024-12-02T17:26:38,319 WARN [Thread-951 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:38,320 WARN [Thread-951 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741869_1052 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,320 WARN [Thread-951 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741869_1052 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:38,320 WARN [Thread-951 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741869_1052 2024-12-02T17:26:38,321 WARN [Thread-951 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:38,322 WARN [IPC Server handler 1 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:38,322 WARN [IPC Server handler 1 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:38,322 WARN [IPC Server handler 1 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:38,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741870_1053 (size=6027) 2024-12-02T17:26:38,725 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.35 KB at sequenceid=45 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/b42aac12c04a45d08a8fbcb70def59e1 2024-12-02T17:26:38,732 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/b42aac12c04a45d08a8fbcb70def59e1 as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/b42aac12c04a45d08a8fbcb70def59e1 2024-12-02T17:26:38,737 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/b42aac12c04a45d08a8fbcb70def59e1, entries=1, sequenceid=45, filesize=5.9 K 2024-12-02T17:26:38,738 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.35 KB/7525, heapSize ~8.11 KB/8304, currentSize=2.10 KB/2150 for 90038ff24035c0472c575f3364061f59 in 431ms, sequenceid=45, compaction requested=false 2024-12-02T17:26:38,738 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 90038ff24035c0472c575f3364061f59: 2024-12-02T17:26:38,739 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=23.5 K, sizeToCheck=16.0 K 2024-12-02T17:26:38,739 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:38,739 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6 because midkey is the same as first or last row 2024-12-02T17:26:38,749 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@2882550b[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741840_1023 to 127.0.0.1:36611 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:38,749 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@a350071[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741860_1043 to 127.0.0.1:36611 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:38,831 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,831 WARN [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(529): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]] 2024-12-02T17:26:38,832 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C37967%2C1733160377246:(num 1733160396817) roll requested 2024-12-02T17:26:38,832 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37967%2C1733160377246.1733160398832 2024-12-02T17:26:38,834 WARN [Thread-956 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741871_1054 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,834 WARN [Thread-956 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741871_1054 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:38,834 WARN [Thread-956 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741871_1054 2024-12-02T17:26:38,835 WARN [Thread-956 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:38,836 WARN [Thread-956 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741872_1055 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,836 WARN [Thread-956 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741872_1055 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:38,836 WARN [Thread-956 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741872_1055 2024-12-02T17:26:38,836 WARN [Thread-956 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:38,837 WARN [Thread-956 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741873_1056 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,837 WARN [Thread-956 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741873_1056 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:38,838 WARN [Thread-956 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741873_1056 2024-12-02T17:26:38,838 WARN [Thread-956 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:38,840 WARN [Thread-956 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741874_1057 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:39625 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:38,840 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49854 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741874_1057] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741874_1057 to mirror 127.0.0.1:39625 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:38,840 WARN [Thread-956 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741874_1057 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:38,840 WARN [Thread-956 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741874_1057 2024-12-02T17:26:38,840 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49854 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741874_1057] {}] datanode.BlockReceiver(316): Block 1073741874 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T17:26:38,840 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49854 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741874_1057] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49854 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:38,841 WARN [Thread-956 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:38,841 WARN [IPC Server handler 1 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:38,841 WARN [IPC Server handler 1 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:38,841 WARN [IPC Server handler 1 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:38,844 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:38,844 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:38,844 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:38,844 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:38,844 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:38,844 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160396817 with entries=15, filesize=13.26 KB; new WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160398832 2024-12-02T17:26:38,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741855_1038 (size=13591) 2024-12-02T17:26:38,849 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40409:40409)] 2024-12-02T17:26:38,849 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 is not closed yet, will try archiving it next time 2024-12-02T17:26:38,849 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160396817 is not closed yet, will try archiving it next time 2024-12-02T17:26:38,849 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160392809 to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs/3b3375d1a925%2C37967%2C1733160377246.1733160392809 2024-12-02T17:26:39,247 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 is not closed yet, will try archiving it next time 2024-12-02T17:26:39,347 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:39,723 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8855): Flush requested on 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:39,723 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 90038ff24035c0472c575f3364061f59 1/1 column families, dataSize=7.35 KB heapSize=8.13 KB 2024-12-02T17:26:39,728 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/1160db26dc46496ca624f90833021dea is 1079, key is tmprow/info:/1733160399722/Put/seqid=0 2024-12-02T17:26:39,730 WARN [Thread-961 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741876_1059 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:39,730 WARN [Thread-961 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741876_1059 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:39,730 WARN [Thread-961 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741876_1059 2024-12-02T17:26:39,731 WARN [Thread-961 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:39,732 WARN [Thread-961 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741877_1060 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:39,732 WARN [Thread-961 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741877_1060 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:39,732 WARN [Thread-961 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741877_1060 2024-12-02T17:26:39,732 WARN [Thread-961 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:39,734 WARN [Thread-961 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741878_1061 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:39,734 WARN [Thread-961 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741878_1061 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:39,734 WARN [Thread-961 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741878_1061 2024-12-02T17:26:39,734 WARN [Thread-961 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:39,735 WARN [Thread-961 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741879_1062 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:39,735 WARN [Thread-961 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741879_1062 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:39,735 WARN [Thread-961 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741879_1062 2024-12-02T17:26:39,736 WARN [Thread-961 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:39,736 WARN [IPC Server handler 3 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:39,736 WARN [IPC Server handler 3 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:39,737 WARN [IPC Server handler 3 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:39,739 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741880_1063 (size=6027) 2024-12-02T17:26:40,067 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:40,140 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.35 KB at sequenceid=55 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/1160db26dc46496ca624f90833021dea 2024-12-02T17:26:40,147 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/1160db26dc46496ca624f90833021dea as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/1160db26dc46496ca624f90833021dea 2024-12-02T17:26:40,153 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/1160db26dc46496ca624f90833021dea, entries=1, sequenceid=55, filesize=5.9 K 2024-12-02T17:26:40,154 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.35 KB/7525, heapSize ~8.11 KB/8304, currentSize=2.10 KB/2150 for 90038ff24035c0472c575f3364061f59 in 431ms, sequenceid=55, compaction requested=true 2024-12-02T17:26:40,154 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 90038ff24035c0472c575f3364061f59: 2024-12-02T17:26:40,154 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=29.3 K, sizeToCheck=16.0 K 2024-12-02T17:26:40,154 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:40,154 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6 because midkey is the same as first or last row 2024-12-02T17:26:40,154 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 90038ff24035c0472c575f3364061f59:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:26:40,154 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:26:40,154 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:26:40,155 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 30048 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:26:40,156 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HStore(1541): 90038ff24035c0472c575f3364061f59/info is initiating minor compaction (all files) 2024-12-02T17:26:40,156 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of 90038ff24035c0472c575f3364061f59/info in TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:40,156 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/b42aac12c04a45d08a8fbcb70def59e1, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/1160db26dc46496ca624f90833021dea] into tmpdir=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp, totalSize=29.3 K 2024-12-02T17:26:40,156 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.Compactor(225): Compacting c1ef748d5d91496b8e6e5262bb56d0b6, keycount=12, bloomtype=ROW, size=17.6 K, encoding=NONE, compression=NONE, seqNum=34, earliestPutTs=1733160390821 2024-12-02T17:26:40,157 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.Compactor(225): Compacting b42aac12c04a45d08a8fbcb70def59e1, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=45, earliestPutTs=1733160398306 2024-12-02T17:26:40,157 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] compactions.Compactor(225): Compacting 1160db26dc46496ca624f90833021dea, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=55, earliestPutTs=1733160399722 2024-12-02T17:26:40,174 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 90038ff24035c0472c575f3364061f59#info#compaction#24 average throughput is 6.16 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:26:40,174 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/5c3b5625b5c3413ba78005da41aaa8f3 is 1080, key is row0002/info:/1733160390821/Put/seqid=0 2024-12-02T17:26:40,176 WARN [Thread-965 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741881_1064 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:40,177 WARN [Thread-965 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741881_1064 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]) is bad. 2024-12-02T17:26:40,177 WARN [Thread-965 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741881_1064 2024-12-02T17:26:40,177 WARN [Thread-965 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36611,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK] 2024-12-02T17:26:40,179 WARN [Thread-965 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741882_1065 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:40,179 WARN [Thread-965 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741882_1065 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:40,179 WARN [Thread-965 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741882_1065 2024-12-02T17:26:40,179 WARN [Thread-965 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:40,181 WARN [Thread-965 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741883_1066 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:40,181 WARN [Thread-965 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741883_1066 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:40,181 WARN [Thread-965 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741883_1066 2024-12-02T17:26:40,181 WARN [Thread-965 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:40,189 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49882 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741884_1067] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741884_1067 to mirror 127.0.0.1:39625 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:40,189 WARN [Thread-965 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741884_1067 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:39625 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:40,189 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49882 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741884_1067] {}] datanode.BlockReceiver(316): Block 1073741884 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T17:26:40,189 WARN [Thread-965 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741884_1067 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:40,189 WARN [Thread-965 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741884_1067 2024-12-02T17:26:40,189 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:49882 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741884_1067] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49882 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:40,190 WARN [Thread-965 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:40,191 WARN [IPC Server handler 1 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T17:26:40,191 WARN [IPC Server handler 1 on default port 41131 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T17:26:40,191 WARN [IPC Server handler 1 on default port 41131 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T17:26:40,194 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741885_1068 (size=18097) 2024-12-02T17:26:40,602 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/5c3b5625b5c3413ba78005da41aaa8f3 as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/5c3b5625b5c3413ba78005da41aaa8f3 2024-12-02T17:26:40,613 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in 90038ff24035c0472c575f3364061f59/info of 90038ff24035c0472c575f3364061f59 into 5c3b5625b5c3413ba78005da41aaa8f3(size=17.7 K), total size for store is 17.7 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:26:40,613 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for 90038ff24035c0472c575f3364061f59: 2024-12-02T17:26:40,613 INFO [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59., storeName=90038ff24035c0472c575f3364061f59/info, priority=13, startTime=1733160400154; duration=0sec 2024-12-02T17:26:40,613 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=17.7 K, sizeToCheck=16.0 K 2024-12-02T17:26:40,613 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:40,613 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/5c3b5625b5c3413ba78005da41aaa8f3 because midkey is the same as first or last row 2024-12-02T17:26:40,613 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=17.7 K, sizeToCheck=16.0 K 2024-12-02T17:26:40,613 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:40,613 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/5c3b5625b5c3413ba78005da41aaa8f3 because midkey is the same as first or last row 2024-12-02T17:26:40,614 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=17.7 K, sizeToCheck=16.0 K 2024-12-02T17:26:40,614 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:40,614 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/5c3b5625b5c3413ba78005da41aaa8f3 because midkey is the same as first or last row 2024-12-02T17:26:40,614 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:26:40,614 DEBUG [RS:0;3b3375d1a925:37967-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 90038ff24035c0472c575f3364061f59:info 2024-12-02T17:26:40,750 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@2882550b[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741870_1053 to 127.0.0.1:45809 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:40,750 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@a350071[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741865_1048 to 127.0.0.1:45891 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:40,849 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:40,850 WARN [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(539): Too many consecutive RollWriter requests, it's a sign of the total number of live datanodes is lower than the tolerable replicas. 2024-12-02T17:26:40,948 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:40,952 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:40,952 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:40,952 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:40,952 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:26:40,953 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@13c2f5a4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:40,953 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1191c470{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:41,069 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4d95479b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/java.io.tmpdir/jetty-localhost-37143-hadoop-hdfs-3_4_1-tests_jar-_-any-12117426715846163029/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:41,069 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@469c1f5d{HTTP/1.1, (http/1.1)}{localhost:37143} 2024-12-02T17:26:41,069 INFO [Time-limited test {}] server.Server(415): Started @128741ms 2024-12-02T17:26:41,070 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:41,164 WARN [Thread-984 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:41,172 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa4e00a79d15e27a9 with lease ID 0x32d3e7d6b46bf24f: from storage DS-ee527e74-ad21-4330-b74a-ac6397764fd4 node DatanodeRegistration(127.0.0.1:39873, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=41459, infoSecurePort=0, ipcPort=46443, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 6, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:41,173 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa4e00a79d15e27a9 with lease ID 0x32d3e7d6b46bf24f: from storage DS-3c29e8b5-6830-4a5e-9080-c2081db2d1db node DatanodeRegistration(127.0.0.1:39873, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=41459, infoSecurePort=0, ipcPort=46443, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:41,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741833_1020 (size=455) 2024-12-02T17:26:41,347 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:41,750 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@2882550b[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741880_1063 to 127.0.0.1:45891 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:41,753 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741855_1038 (size=13591) 2024-12-02T17:26:41,823 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160377674 to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs/3b3375d1a925%2C37967%2C1733160377246.1733160377674 2024-12-02T17:26:41,825 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160396817 to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs/3b3375d1a925%2C37967%2C1733160377246.1733160396817 2024-12-02T17:26:42,067 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:42,850 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:43,348 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:43,752 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741885_1068 (size=18097) 2024-12-02T17:26:44,067 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:44,168 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@4ea866cc[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:39873, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=41459, infoSecurePort=0, ipcPort=46443, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741833_1020 to 127.0.0.1:45809 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:44,850 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:45,348 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:46,068 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:46,850 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:47,168 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:26:47,349 INFO [master:store-WAL-Roller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:47,474 ERROR [FSHLog-0-hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData-prefix:3b3375d1a925,45913,1733160377186 {}] wal.AbstractFSWAL(1838): appendAndSync throws IOException. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:47,474 WARN [FSHLog-0-hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData-prefix:3b3375d1a925,45913,1733160377186 {}] wal.AbstractFSWAL(2174): append entry failed java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:47,474 DEBUG [master:store-WAL-Roller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C45913%2C1733160377186:(num 1733160377372) roll requested 2024-12-02T17:26:47,475 INFO [master:store-WAL-Roller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C45913%2C1733160377186.1733160407474 2024-12-02T17:26:47,477 WARN [Thread-1007 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741886_1069 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:47,478 WARN [Thread-1007 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741886_1069 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:39873,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:47,478 WARN [Thread-1007 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741886_1069 2024-12-02T17:26:47,478 WARN [Thread-1007 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:47,482 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:47,482 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:47,482 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:47,482 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:47,482 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:47,483 INFO [master:store-WAL-Roller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 with entries=54, filesize=26.67 KB; new WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160407474 2024-12-02T17:26:47,483 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:47,483 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:47,483 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 2024-12-02T17:26:47,483 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41459:41459),(127.0.0.1/127.0.0.1:40409:40409)] 2024-12-02T17:26:47,483 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 is not closed yet, will try archiving it next time 2024-12-02T17:26:47,484 WARN [IPC Server handler 2 on default port 41131 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 has not been closed. Lease recovery is in progress. RecoveryId = 1071 for block blk_1073741830_1006 2024-12-02T17:26:47,484 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 after 1ms 2024-12-02T17:26:48,068 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:48,851 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:50,068 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:50,169 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@4ea866cc[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:39873, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=41459, infoSecurePort=0, ipcPort=46443, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741830_1071 to 127.0.0.1:45809 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:50,851 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:51,485 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/WALs/3b3375d1a925,45913,1733160377186/3b3375d1a925%2C45913%2C1733160377186.1733160377372 after 4002ms 2024-12-02T17:26:52,069 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:52,851 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,069 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,670 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37967%2C1733160377246.1733160414669 2024-12-02T17:26:54,672 WARN [Thread-1015 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741888_1072 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,673 WARN [Thread-1015 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741888_1072 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:54,673 WARN [Thread-1015 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741888_1072 2024-12-02T17:26:54,673 WARN [Thread-1015 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:54,676 WARN [Thread-1015 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741889_1073 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:45809 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,675 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1129077784_22 at /127.0.0.1:43786 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741889_1073] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741889_1073 to mirror 127.0.0.1:45809 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:54,676 WARN [Thread-1015 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741889_1073 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:54,676 WARN [Thread-1015 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741889_1073 2024-12-02T17:26:54,676 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1129077784_22 at /127.0.0.1:43786 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741889_1073] {}] datanode.BlockReceiver(316): Block 1073741889 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T17:26:54,676 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1129077784_22 at /127.0.0.1:43786 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741889_1073] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:43786 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:54,676 WARN [Thread-1015 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:54,680 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,680 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,680 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,681 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,681 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,681 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160398832 with entries=13, filesize=12.60 KB; new WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160414669 2024-12-02T17:26:54,682 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41459:41459),(127.0.0.1/127.0.0.1:40409:40409)] 2024-12-02T17:26:54,682 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160398832 is not closed yet, will try archiving it next time 2024-12-02T17:26:54,682 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741875_1058 (size=12911) 2024-12-02T17:26:54,687 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8855): Flush requested on 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:54,687 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 90038ff24035c0472c575f3364061f59 1/1 column families, dataSize=7.35 KB heapSize=8.13 KB 2024-12-02T17:26:54,696 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/0e1e3e9edf4e4b2d9950ab495d207050 is 1080, key is row0013/info:/1733160414683/Put/seqid=0 2024-12-02T17:26:54,698 WARN [Thread-1022 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741891_1075 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,698 WARN [Thread-1022 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741891_1075 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:54,698 WARN [Thread-1022 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741891_1075 2024-12-02T17:26:54,699 WARN [Thread-1022 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:54,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741892_1076 (size=8190) 2024-12-02T17:26:54,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741892_1076 (size=8190) 2024-12-02T17:26:54,710 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.35 KB at sequenceid=66 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/0e1e3e9edf4e4b2d9950ab495d207050 2024-12-02T17:26:54,716 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/0e1e3e9edf4e4b2d9950ab495d207050 as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/0e1e3e9edf4e4b2d9950ab495d207050 2024-12-02T17:26:54,723 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/0e1e3e9edf4e4b2d9950ab495d207050, entries=3, sequenceid=66, filesize=8.0 K 2024-12-02T17:26:54,724 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.35 KB/7527, heapSize ~8.11 KB/8304, currentSize=9.46 KB/9683 for 90038ff24035c0472c575f3364061f59 in 37ms, sequenceid=66, compaction requested=false 2024-12-02T17:26:54,724 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 90038ff24035c0472c575f3364061f59: 2024-12-02T17:26:54,724 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=25.7 K, sizeToCheck=16.0 K 2024-12-02T17:26:54,724 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:26:54,724 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/5c3b5625b5c3413ba78005da41aaa8f3 because midkey is the same as first or last row 2024-12-02T17:26:54,852 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(580): java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,852 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.FSHLog(556): LowReplication-Roller was enabled. 2024-12-02T17:26:54,906 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T17:26:54,906 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:26:54,906 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:54,906 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:54,907 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:54,907 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T17:26:54,907 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T17:26:54,907 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=128730042, stopped=false 2024-12-02T17:26:54,907 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=3b3375d1a925,45913,1733160377186 2024-12-02T17:26:54,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:54,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:54,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:26:54,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:54,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:54,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:54,911 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:26:54,911 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:26:54,912 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:54,912 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:54,912 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:26:54,912 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:54,912 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:54,912 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '3b3375d1a925,37967,1733160377246' ***** 2024-12-02T17:26:54,912 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T17:26:54,912 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '3b3375d1a925,37629,1733160378005' ***** 2024-12-02T17:26:54,912 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T17:26:54,913 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T17:26:54,913 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T17:26:54,913 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T17:26:54,913 INFO [RS:1;3b3375d1a925:37629 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T17:26:54,913 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T17:26:54,913 INFO [RS:1;3b3375d1a925:37629 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T17:26:54,913 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(959): stopping server 3b3375d1a925,37629,1733160378005 2024-12-02T17:26:54,913 INFO [RS:0;3b3375d1a925:37967 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T17:26:54,913 INFO [RS:1;3b3375d1a925:37629 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:26:54,913 INFO [RS:0;3b3375d1a925:37967 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T17:26:54,913 INFO [RS:1;3b3375d1a925:37629 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;3b3375d1a925:37629. 2024-12-02T17:26:54,913 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(3091): Received CLOSE for 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:54,913 DEBUG [RS:1;3b3375d1a925:37629 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:54,913 DEBUG [RS:1;3b3375d1a925:37629 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:54,913 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(976): stopping server 3b3375d1a925,37629,1733160378005; all regions closed. 2024-12-02T17:26:54,913 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(959): stopping server 3b3375d1a925,37967,1733160377246 2024-12-02T17:26:54,913 INFO [RS:0;3b3375d1a925:37967 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:26:54,914 INFO [RS:0;3b3375d1a925:37967 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;3b3375d1a925:37967. 2024-12-02T17:26:54,914 DEBUG [RS:0;3b3375d1a925:37967 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:26:54,914 DEBUG [RS:0;3b3375d1a925:37967 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:54,914 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,914 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T17:26:54,914 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,914 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 90038ff24035c0472c575f3364061f59, disabling compactions & flushes 2024-12-02T17:26:54,914 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T17:26:54,914 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T17:26:54,914 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,914 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:54,914 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T17:26:54,914 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:54,914 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,914 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. after waiting 0 ms 2024-12-02T17:26:54,914 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:54,914 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,914 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-02T17:26:54,914 DEBUG [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1325): Online Regions={90038ff24035c0472c575f3364061f59=TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59., 1588230740=hbase:meta,,1.1588230740} 2024-12-02T17:26:54,914 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2902): Flushing 90038ff24035c0472c575f3364061f59 1/1 column families, dataSize=9.46 KB heapSize=10.38 KB 2024-12-02T17:26:54,914 DEBUG [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, 90038ff24035c0472c575f3364061f59 2024-12-02T17:26:54,915 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:26:54,915 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,915 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:26:54,915 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:26:54,915 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:26:54,915 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,915 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:26:54,915 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005/3b3375d1a925%2C37629%2C1733160378005.1733160378229 2024-12-02T17:26:54,915 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.71 KB heapSize=3.75 KB 2024-12-02T17:26:54,915 WARN [IPC Server handler 1 on default port 41131 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005/3b3375d1a925%2C37629%2C1733160378005.1733160378229 has not been closed. Lease recovery is in progress. RecoveryId = 1077 for block blk_1073741837_1013 2024-12-02T17:26:54,915 ERROR [FSHLog-0-hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca-prefix:3b3375d1a925,37967,1733160377246.meta {}] wal.AbstractFSWAL(1838): appendAndSync throws IOException. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,916 WARN [FSHLog-0-hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca-prefix:3b3375d1a925,37967,1733160377246.meta {}] wal.AbstractFSWAL(2174): append entry failed java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,916 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005/3b3375d1a925%2C37629%2C1733160378005.1733160378229 after 1ms 2024-12-02T17:26:54,916 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C37967%2C1733160377246.meta:.meta(num 1733160377861) roll requested 2024-12-02T17:26:54,916 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C37967%2C1733160377246.meta.1733160414916.meta 2024-12-02T17:26:54,918 WARN [Thread-1030 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741893_1078 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,919 WARN [Thread-1030 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741893_1078 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:39873,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:54,919 WARN [Thread-1030 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741893_1078 2024-12-02T17:26:54,919 WARN [Thread-1030 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:54,920 WARN [Thread-1030 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741894_1079 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,921 WARN [Thread-1030 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741894_1079 in pipeline [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK], DatanodeInfoWithStorage[127.0.0.1:39873,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]) is bad. 2024-12-02T17:26:54,921 WARN [Thread-1030 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741894_1079 2024-12-02T17:26:54,921 WARN [Thread-1030 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK] 2024-12-02T17:26:54,921 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/72b97a6763d44d00a5658fbeb27ae2a5 is 1080, key is row0015/info:/1733160414688/Put/seqid=0 2024-12-02T17:26:54,924 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:43826 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741896_1081] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10]'}, localName='127.0.0.1:34399', datanodeUuid='16a14b74-c77d-4c52-aef6-a3e479cabf82', xmitsInProgress=0}:Exception transferring block BP-797259812-172.17.0.3-1733160376449:blk_1073741896_1081 to mirror 127.0.0.1:45809 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:54,924 WARN [Thread-1031 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741896_1081 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:45809 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,924 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:43826 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741896_1081] {}] datanode.BlockReceiver(316): Block 1073741896 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T17:26:54,924 WARN [Thread-1031 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741896_1081 in pipeline [DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK], DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:54,924 WARN [Thread-1031 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741896_1081 2024-12-02T17:26:54,924 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1804418420_22 at /127.0.0.1:43826 [Receiving block BP-797259812-172.17.0.3-1733160376449:blk_1073741896_1081] {}] datanode.DataXceiver(331): 127.0.0.1:34399:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:43826 dst: /127.0.0.1:34399 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:54,925 WARN [Thread-1031 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:54,925 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,925 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,925 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,925 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,925 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:54,926 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta with entries=8, filesize=2.33 KB; new WAL /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160414916.meta 2024-12-02T17:26:54,926 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,926 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:39625,DS-0338ef0e-e232-4469-8f02-5fdc748ca069,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,926 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta 2024-12-02T17:26:54,926 WARN [Thread-1031 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741897_1082 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,926 WARN [Thread-1031 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741897_1082 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:39873,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:54,926 WARN [Thread-1031 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741897_1082 2024-12-02T17:26:54,926 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40409:40409),(127.0.0.1/127.0.0.1:41459:41459)] 2024-12-02T17:26:54,926 WARN [IPC Server handler 1 on default port 41131 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta has not been closed. Lease recovery is in progress. RecoveryId = 1083 for block blk_1073741834_1010 2024-12-02T17:26:54,927 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta is not closed yet, will try archiving it next time 2024-12-02T17:26:54,927 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta after 1ms 2024-12-02T17:26:54,927 WARN [Thread-1031 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:54,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741898_1084 (size=14660) 2024-12-02T17:26:54,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741898_1084 (size=14660) 2024-12-02T17:26:54,931 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=9.46 KB at sequenceid=78 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/72b97a6763d44d00a5658fbeb27ae2a5 2024-12-02T17:26:54,938 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/.tmp/info/72b97a6763d44d00a5658fbeb27ae2a5 as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/72b97a6763d44d00a5658fbeb27ae2a5 2024-12-02T17:26:54,943 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/72b97a6763d44d00a5658fbeb27ae2a5, entries=9, sequenceid=78, filesize=14.3 K 2024-12-02T17:26:54,944 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3140): Finished flush of dataSize ~9.46 KB/9683, heapSize ~10.36 KB/10608, currentSize=0 B/0 for 90038ff24035c0472c575f3364061f59 in 30ms, sequenceid=78, compaction requested=true 2024-12-02T17:26:54,945 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] regionserver.HStore(2317): Moving the files [hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/d7340ed25fd247928cf6d5048618e1db, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/234fe182545e476cb5e8e763fe9b490a, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/517367aec22d42aea6abc75b1cbbc5dd, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/b42aac12c04a45d08a8fbcb70def59e1, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/1160db26dc46496ca624f90833021dea] to archive 2024-12-02T17:26:54,946 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T17:26:54,949 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/info/5189157dbf304022a6984653f96ce445 is 203, key is TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59./info:regioninfo/1733160378489/Put/seqid=0 2024-12-02T17:26:54,950 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/d7340ed25fd247928cf6d5048618e1db to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/archive/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/d7340ed25fd247928cf6d5048618e1db 2024-12-02T17:26:54,951 WARN [Thread-1043 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741899_1085 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,951 WARN [Thread-1043 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741899_1085 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:34399,DS-73e19a57-93ad-4448-abc6-2e0fa5f1af70,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:54,951 WARN [Thread-1043 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741899_1085 2024-12-02T17:26:54,951 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/234fe182545e476cb5e8e763fe9b490a to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/archive/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/234fe182545e476cb5e8e763fe9b490a 2024-12-02T17:26:54,952 WARN [Thread-1043 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:54,953 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6 to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/archive/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/c1ef748d5d91496b8e6e5262bb56d0b6 2024-12-02T17:26:54,953 WARN [Thread-1043 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741900_1086 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,953 WARN [Thread-1043 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741900_1086 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:39873,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:54,953 WARN [Thread-1043 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741900_1086 2024-12-02T17:26:54,954 WARN [Thread-1043 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:54,955 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/517367aec22d42aea6abc75b1cbbc5dd to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/archive/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/517367aec22d42aea6abc75b1cbbc5dd 2024-12-02T17:26:54,956 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/b42aac12c04a45d08a8fbcb70def59e1 to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/archive/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/b42aac12c04a45d08a8fbcb70def59e1 2024-12-02T17:26:54,958 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/1160db26dc46496ca624f90833021dea to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/archive/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/info/1160db26dc46496ca624f90833021dea 2024-12-02T17:26:54,958 DEBUG [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] regionserver.HRegionServer(3532): Failed to report file archival(s) to Master. This will be retried. org.apache.hadoop.hbase.ipc.StoppedRpcClientException: Call to address=3b3375d1a925:45913 failed on local exception: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.hbase.ipc.IPCUtil.wrapException(IPCUtil.java:237) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:395) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:430) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:425) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:117) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.setException(Call.java:132) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:451) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:336) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:606) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$BlockingStub.reportFileArchival(RegionServerStatusProtos.java:17350) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegionServer.reportFileArchivalForQuotas(HRegionServer.java:3516) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.reportArchivedFilesForQuota(HStore.java:2412) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.removeCompactedfiles(HStore.java:2347) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.closeWithoutLock(HStore.java:738) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.close(HStore.java:804) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion$2.call(HRegion.java:1912) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion$2.call(HRegion.java:1909) ~[classes/:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) ~[?:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at org.apache.hadoop.hbase.ipc.AbstractRpcClient.getConnection(AbstractRpcClient.java:366) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:448) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] ... 16 more 2024-12-02T17:26:54,959 WARN [StoreCloser-TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59.-1 {}] regionserver.HStore(2414): Failed to report archival of files: [d7340ed25fd247928cf6d5048618e1db=10347, 234fe182545e476cb5e8e763fe9b490a=12506, c1ef748d5d91496b8e6e5262bb56d0b6=17994, 517367aec22d42aea6abc75b1cbbc5dd=6027, b42aac12c04a45d08a8fbcb70def59e1=6027, 1160db26dc46496ca624f90833021dea=6027] 2024-12-02T17:26:54,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741901_1087 (size=7089) 2024-12-02T17:26:54,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741901_1087 (size=7089) 2024-12-02T17:26:54,969 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.50 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/info/5189157dbf304022a6984653f96ce445 2024-12-02T17:26:54,973 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/default/TestLogRolling-testLogRollOnDatanodeDeath/90038ff24035c0472c575f3364061f59/recovered.edits/81.seqid, newMaxSeqId=81, maxSeqId=1 2024-12-02T17:26:54,973 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:54,973 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 90038ff24035c0472c575f3364061f59: Waiting for close lock at 1733160414914Running coprocessor pre-close hooks at 1733160414914Disabling compacts and flushes for region at 1733160414914Disabling writes for close at 1733160414914Obtaining lock to block concurrent updates at 1733160414914Preparing flush snapshotting stores in 90038ff24035c0472c575f3364061f59 at 1733160414914Finished memstore snapshotting TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59., syncing WAL and waiting on mvcc, flushsize=dataSize=9683, getHeapSize=10608, getOffHeapSize=0, getCellsCount=9 at 1733160414915 (+1 ms)Flushing stores of TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. at 1733160414916 (+1 ms)Flushing 90038ff24035c0472c575f3364061f59/info: creating writer at 1733160414916Flushing 90038ff24035c0472c575f3364061f59/info: appending metadata at 1733160414921 (+5 ms)Flushing 90038ff24035c0472c575f3364061f59/info: closing flushed file at 1733160414921Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2ea1f80e: reopening flushed file at 1733160414937 (+16 ms)Finished flush of dataSize ~9.46 KB/9683, heapSize ~10.36 KB/10608, currentSize=0 B/0 for 90038ff24035c0472c575f3364061f59 in 30ms, sequenceid=78, compaction requested=true at 1733160414944 (+7 ms)Writing region close event to WAL at 1733160414969 (+25 ms)Running coprocessor post-close hooks at 1733160414973 (+4 ms)Closed at 1733160414973 2024-12-02T17:26:54,974 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRollOnDatanodeDeath,,1733160378128.90038ff24035c0472c575f3364061f59. 2024-12-02T17:26:54,989 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/ns/ff409ab8eae344cf993c6de7e369e86a is 43, key is default/ns:d/1733160377912/Put/seqid=0 2024-12-02T17:26:54,990 WARN [Thread-1050 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741902_1088 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,990 WARN [Thread-1050 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741902_1088 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK], DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK]) is bad. 2024-12-02T17:26:54,990 WARN [Thread-1050 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741902_1088 2024-12-02T17:26:54,991 WARN [Thread-1050 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45809,DS-305215bb-4b01-4aa8-8b61-3f08c0cb0b6b,DISK] 2024-12-02T17:26:54,992 WARN [Thread-1050 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741903_1089 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:26:54,992 WARN [Thread-1050 {}] hdfs.DataStreamer(1731): Error Recovery for BP-797259812-172.17.0.3-1733160376449:blk_1073741903_1089 in pipeline [DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK], DatanodeInfoWithStorage[127.0.0.1:39873,DS-ee527e74-ad21-4330-b74a-ac6397764fd4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK]) is bad. 2024-12-02T17:26:54,992 WARN [Thread-1050 {}] hdfs.DataStreamer(1850): Abandoning BP-797259812-172.17.0.3-1733160376449:blk_1073741903_1089 2024-12-02T17:26:54,992 WARN [Thread-1050 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:45891,DS-176dfedc-f5b8-4b28-b862-d6e648f52d33,DISK] 2024-12-02T17:26:54,996 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741904_1090 (size=5153) 2024-12-02T17:26:54,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741904_1090 (size=5153) 2024-12-02T17:26:54,997 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/ns/ff409ab8eae344cf993c6de7e369e86a 2024-12-02T17:26:55,016 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/table/a742305aed014010a635ba8d933c395c is 77, key is TestLogRolling-testLogRollOnDatanodeDeath/table:state/1733160378501/Put/seqid=0 2024-12-02T17:26:55,021 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741905_1091 (size=5424) 2024-12-02T17:26:55,021 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741905_1091 (size=5424) 2024-12-02T17:26:55,022 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=146 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/table/a742305aed014010a635ba8d933c395c 2024-12-02T17:26:55,027 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/info/5189157dbf304022a6984653f96ce445 as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/info/5189157dbf304022a6984653f96ce445 2024-12-02T17:26:55,032 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/info/5189157dbf304022a6984653f96ce445, entries=10, sequenceid=11, filesize=6.9 K 2024-12-02T17:26:55,033 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/ns/ff409ab8eae344cf993c6de7e369e86a as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/ns/ff409ab8eae344cf993c6de7e369e86a 2024-12-02T17:26:55,039 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/ns/ff409ab8eae344cf993c6de7e369e86a, entries=2, sequenceid=11, filesize=5.0 K 2024-12-02T17:26:55,040 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/.tmp/table/a742305aed014010a635ba8d933c395c as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/table/a742305aed014010a635ba8d933c395c 2024-12-02T17:26:55,045 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/table/a742305aed014010a635ba8d933c395c, entries=2, sequenceid=11, filesize=5.3 K 2024-12-02T17:26:55,046 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.71 KB/1752, heapSize ~3.45 KB/3536, currentSize=0 B/0 for 1588230740 in 131ms, sequenceid=11, compaction requested=false 2024-12-02T17:26:55,050 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-02T17:26:55,051 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:26:55,051 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:26:55,051 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160414915Running coprocessor pre-close hooks at 1733160414915Disabling compacts and flushes for region at 1733160414915Disabling writes for close at 1733160414915Obtaining lock to block concurrent updates at 1733160414915Preparing flush snapshotting stores in 1588230740 at 1733160414915Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1752, getHeapSize=3776, getOffHeapSize=0, getCellsCount=14 at 1733160414915Flushing stores of hbase:meta,,1.1588230740 at 1733160414927 (+12 ms)Flushing 1588230740/info: creating writer at 1733160414927Flushing 1588230740/info: appending metadata at 1733160414949 (+22 ms)Flushing 1588230740/info: closing flushed file at 1733160414949Flushing 1588230740/ns: creating writer at 1733160414974 (+25 ms)Flushing 1588230740/ns: appending metadata at 1733160414988 (+14 ms)Flushing 1588230740/ns: closing flushed file at 1733160414988Flushing 1588230740/table: creating writer at 1733160415002 (+14 ms)Flushing 1588230740/table: appending metadata at 1733160415016 (+14 ms)Flushing 1588230740/table: closing flushed file at 1733160415016Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@76f6ac23: reopening flushed file at 1733160415026 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1ad77635: reopening flushed file at 1733160415033 (+7 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@45bea647: reopening flushed file at 1733160415039 (+6 ms)Finished flush of dataSize ~1.71 KB/1752, heapSize ~3.45 KB/3536, currentSize=0 B/0 for 1588230740 in 131ms, sequenceid=11, compaction requested=false at 1733160415046 (+7 ms)Writing region close event to WAL at 1733160415047 (+1 ms)Running coprocessor post-close hooks at 1733160415051 (+4 ms)Closed at 1733160415051 2024-12-02T17:26:55,051 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T17:26:55,065 INFO [regionserver/3b3375d1a925:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-02T17:26:55,065 INFO [regionserver/3b3375d1a925:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-02T17:26:55,083 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.1733160398832 to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs/3b3375d1a925%2C37967%2C1733160377246.1733160398832 2024-12-02T17:26:55,115 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(976): stopping server 3b3375d1a925,37967,1733160377246; all regions closed. 2024-12-02T17:26:55,115 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:55,115 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:55,115 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:55,115 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:55,115 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:55,117 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741895_1080 (size=825) 2024-12-02T17:26:55,117 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741895_1080 (size=825) 2024-12-02T17:26:55,538 INFO [regionserver/3b3375d1a925:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:26:55,544 INFO [regionserver/3b3375d1a925:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-02T17:26:55,544 INFO [regionserver/3b3375d1a925:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-02T17:26:55,751 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@a350071[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:34399, datanodeUuid=16a14b74-c77d-4c52-aef6-a3e479cabf82, infoPort=40409, infoSecurePort=0, ipcPort=36229, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741875_1058 to 127.0.0.1:45809 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:56,066 INFO [regionserver/3b3375d1a925:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:26:56,169 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@4ea866cc[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:39873, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=41459, infoSecurePort=0, ipcPort=46443, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741837_1077 to 127.0.0.1:45891 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:56,170 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@7f053175[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:39873, datanodeUuid=9f811f10-21a0-40f9-b54a-2bca063deae0, infoPort=41459, infoSecurePort=0, ipcPort=46443, storageInfo=lv=-57;cid=testClusterID;nsid=545529960;c=1733160376449):Failed to transfer BP-797259812-172.17.0.3-1733160376449:blk_1073741834_1083 to 127.0.0.1:39625 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:26:57,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnDatanodeDeath 2024-12-02T17:26:57,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:26:57,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:26:57,927 INFO [master/3b3375d1a925:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T17:26:57,927 INFO [master/3b3375d1a925:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T17:26:58,916 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37629,1733160378005/3b3375d1a925%2C37629%2C1733160378005.1733160378229 after 4001ms 2024-12-02T17:26:58,918 DEBUG [RS:1;3b3375d1a925:37629 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs 2024-12-02T17:26:58,918 INFO [RS:1;3b3375d1a925:37629 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C37629%2C1733160378005:(num 1733160378229) 2024-12-02T17:26:58,919 DEBUG [RS:1;3b3375d1a925:37629 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:58,919 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:26:58,919 INFO [RS:1;3b3375d1a925:37629 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:26:58,919 INFO [RS:1;3b3375d1a925:37629 {}] hbase.ChoreService(370): Chore service for: regionserver/3b3375d1a925:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-02T17:26:58,919 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T17:26:58,919 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:26:58,919 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T17:26:58,919 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T17:26:58,919 INFO [RS:1;3b3375d1a925:37629 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:26:58,919 INFO [RS:1;3b3375d1a925:37629 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:37629 2024-12-02T17:26:58,921 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:26:58,921 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/3b3375d1a925,37629,1733160378005 2024-12-02T17:26:58,921 INFO [RS:1;3b3375d1a925:37629 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:26:58,923 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [3b3375d1a925,37629,1733160378005] 2024-12-02T17:26:58,924 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/3b3375d1a925,37629,1733160378005 already deleted, retry=false 2024-12-02T17:26:58,924 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 3b3375d1a925,37629,1733160378005 expired; onlineServers=1 2024-12-02T17:26:58,927 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta after 4001ms 2024-12-02T17:26:58,928 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/WALs/3b3375d1a925,37967,1733160377246/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta to hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs/3b3375d1a925%2C37967%2C1733160377246.meta.1733160377861.meta 2024-12-02T17:26:58,930 DEBUG [RS:0;3b3375d1a925:37967 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs 2024-12-02T17:26:58,930 INFO [RS:0;3b3375d1a925:37967 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C37967%2C1733160377246.meta:.meta(num 1733160414916) 2024-12-02T17:26:58,931 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:58,931 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:58,931 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:58,931 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:58,931 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:58,933 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741890_1074 (size=14682) 2024-12-02T17:26:58,933 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741890_1074 (size=14682) 2024-12-02T17:26:58,935 DEBUG [RS:0;3b3375d1a925:37967 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/oldWALs 2024-12-02T17:26:58,935 INFO [RS:0;3b3375d1a925:37967 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C37967%2C1733160377246:(num 1733160414669) 2024-12-02T17:26:58,935 DEBUG [RS:0;3b3375d1a925:37967 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:26:58,935 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:26:58,935 INFO [RS:0;3b3375d1a925:37967 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:26:58,935 INFO [RS:0;3b3375d1a925:37967 {}] hbase.ChoreService(370): Chore service for: regionserver/3b3375d1a925:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-02T17:26:58,936 INFO [RS:0;3b3375d1a925:37967 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:26:58,936 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:26:58,936 INFO [RS:0;3b3375d1a925:37967 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:37967 2024-12-02T17:26:58,938 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/3b3375d1a925,37967,1733160377246 2024-12-02T17:26:58,938 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:26:58,938 INFO [RS:0;3b3375d1a925:37967 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:26:58,940 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [3b3375d1a925,37967,1733160377246] 2024-12-02T17:26:58,941 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/3b3375d1a925,37967,1733160377246 already deleted, retry=false 2024-12-02T17:26:58,941 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 3b3375d1a925,37967,1733160377246 expired; onlineServers=0 2024-12-02T17:26:58,941 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '3b3375d1a925,45913,1733160377186' ***** 2024-12-02T17:26:58,941 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T17:26:58,941 INFO [M:0;3b3375d1a925:45913 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:26:58,941 INFO [M:0;3b3375d1a925:45913 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:26:58,942 DEBUG [M:0;3b3375d1a925:45913 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T17:26:58,942 DEBUG [M:0;3b3375d1a925:45913 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T17:26:58,942 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T17:26:58,942 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160377489 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160377489,5,FailOnTimeoutGroup] 2024-12-02T17:26:58,942 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160377486 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160377486,5,FailOnTimeoutGroup] 2024-12-02T17:26:58,942 INFO [M:0;3b3375d1a925:45913 {}] hbase.ChoreService(370): Chore service for: master/3b3375d1a925:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T17:26:58,942 INFO [M:0;3b3375d1a925:45913 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:26:58,942 DEBUG [M:0;3b3375d1a925:45913 {}] master.HMaster(1795): Stopping service threads 2024-12-02T17:26:58,942 INFO [M:0;3b3375d1a925:45913 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T17:26:58,942 INFO [M:0;3b3375d1a925:45913 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:26:58,942 INFO [M:0;3b3375d1a925:45913 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T17:26:58,943 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T17:26:58,943 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T17:26:58,943 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:26:58,943 DEBUG [M:0;3b3375d1a925:45913 {}] zookeeper.ZKUtil(347): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T17:26:58,943 WARN [M:0;3b3375d1a925:45913 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T17:26:58,944 INFO [M:0;3b3375d1a925:45913 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/.lastflushedseqids 2024-12-02T17:26:58,949 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741906_1092 (size=130) 2024-12-02T17:26:58,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741906_1092 (size=130) 2024-12-02T17:26:58,950 INFO [M:0;3b3375d1a925:45913 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T17:26:58,950 INFO [M:0;3b3375d1a925:45913 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T17:26:58,950 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:26:58,950 INFO [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:58,950 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:58,950 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:26:58,950 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:58,951 INFO [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=23.25 KB heapSize=29.49 KB 2024-12-02T17:26:58,966 DEBUG [M:0;3b3375d1a925:45913 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/dd94c61d373a4298b4755c74800d47c1 is 82, key is hbase:meta,,1/info:regioninfo/1733160377893/Put/seqid=0 2024-12-02T17:26:58,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741907_1093 (size=5672) 2024-12-02T17:26:58,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741907_1093 (size=5672) 2024-12-02T17:26:58,971 INFO [M:0;3b3375d1a925:45913 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=60 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/dd94c61d373a4298b4755c74800d47c1 2024-12-02T17:26:58,998 DEBUG [M:0;3b3375d1a925:45913 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/8374f765fb764fc9b579ce5919165b3a is 774, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733160378507/Put/seqid=0 2024-12-02T17:26:59,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741908_1094 (size=6255) 2024-12-02T17:26:59,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741908_1094 (size=6255) 2024-12-02T17:26:59,005 INFO [M:0;3b3375d1a925:45913 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=22.58 KB at sequenceid=60 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/8374f765fb764fc9b579ce5919165b3a 2024-12-02T17:26:59,011 INFO [M:0;3b3375d1a925:45913 {}] regionserver.StoreFileReader(518): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 8374f765fb764fc9b579ce5919165b3a 2024-12-02T17:26:59,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:59,023 INFO [RS:1;3b3375d1a925:37629 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:26:59,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37629-0x1009c06a5f00002, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:59,023 INFO [RS:1;3b3375d1a925:37629 {}] regionserver.HRegionServer(1031): Exiting; stopping=3b3375d1a925,37629,1733160378005; zookeeper connection closed. 2024-12-02T17:26:59,023 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7b28ebcf {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7b28ebcf 2024-12-02T17:26:59,025 DEBUG [M:0;3b3375d1a925:45913 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7a8f164a97e945d8a910965c3000d0df is 69, key is 3b3375d1a925,37629,1733160378005/rs:state/1733160378051/Put/seqid=0 2024-12-02T17:26:59,030 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741909_1095 (size=5224) 2024-12-02T17:26:59,030 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741909_1095 (size=5224) 2024-12-02T17:26:59,030 INFO [M:0;3b3375d1a925:45913 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=130 B at sequenceid=60 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7a8f164a97e945d8a910965c3000d0df 2024-12-02T17:26:59,040 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:59,040 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1009c06a5f00001, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:59,040 INFO [RS:0;3b3375d1a925:37967 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:26:59,040 INFO [RS:0;3b3375d1a925:37967 {}] regionserver.HRegionServer(1031): Exiting; stopping=3b3375d1a925,37967,1733160377246; zookeeper connection closed. 2024-12-02T17:26:59,040 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@27dbc211 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@27dbc211 2024-12-02T17:26:59,041 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 2 regionserver(s) complete 2024-12-02T17:26:59,055 DEBUG [M:0;3b3375d1a925:45913 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/344dd836a1084928bd50a0ffb308e787 is 52, key is load_balancer_on/state:d/1733160377987/Put/seqid=0 2024-12-02T17:26:59,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741910_1096 (size=5056) 2024-12-02T17:26:59,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741910_1096 (size=5056) 2024-12-02T17:26:59,063 INFO [M:0;3b3375d1a925:45913 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=60 (bloomFilter=true), to=hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/344dd836a1084928bd50a0ffb308e787 2024-12-02T17:26:59,069 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/dd94c61d373a4298b4755c74800d47c1 as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/dd94c61d373a4298b4755c74800d47c1 2024-12-02T17:26:59,075 INFO [M:0;3b3375d1a925:45913 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/dd94c61d373a4298b4755c74800d47c1, entries=8, sequenceid=60, filesize=5.5 K 2024-12-02T17:26:59,076 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/8374f765fb764fc9b579ce5919165b3a as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/8374f765fb764fc9b579ce5919165b3a 2024-12-02T17:26:59,081 INFO [M:0;3b3375d1a925:45913 {}] regionserver.StoreFileReader(518): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 8374f765fb764fc9b579ce5919165b3a 2024-12-02T17:26:59,081 INFO [M:0;3b3375d1a925:45913 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/8374f765fb764fc9b579ce5919165b3a, entries=6, sequenceid=60, filesize=6.1 K 2024-12-02T17:26:59,082 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7a8f164a97e945d8a910965c3000d0df as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/7a8f164a97e945d8a910965c3000d0df 2024-12-02T17:26:59,088 INFO [M:0;3b3375d1a925:45913 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/7a8f164a97e945d8a910965c3000d0df, entries=2, sequenceid=60, filesize=5.1 K 2024-12-02T17:26:59,089 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/344dd836a1084928bd50a0ffb308e787 as hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/344dd836a1084928bd50a0ffb308e787 2024-12-02T17:26:59,094 INFO [M:0;3b3375d1a925:45913 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41131/user/jenkins/test-data/ce76cb12-0146-d205-316b-90869ef4f6ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/344dd836a1084928bd50a0ffb308e787, entries=1, sequenceid=60, filesize=4.9 K 2024-12-02T17:26:59,095 INFO [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(3140): Finished flush of dataSize ~23.25 KB/23805, heapSize ~29.43 KB/30136, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 145ms, sequenceid=60, compaction requested=false 2024-12-02T17:26:59,096 INFO [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:26:59,096 DEBUG [M:0;3b3375d1a925:45913 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160418950Disabling compacts and flushes for region at 1733160418950Disabling writes for close at 1733160418950Obtaining lock to block concurrent updates at 1733160418951 (+1 ms)Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733160418951Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=23805, getHeapSize=30136, getOffHeapSize=0, getCellsCount=71 at 1733160418951Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733160418952 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733160418952Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733160418966 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733160418966Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733160418977 (+11 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733160418997 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733160418997Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733160419011 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733160419025 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733160419025Flushing 1595e783b53d99cd5eef43b6debb2682/state: creating writer at 1733160419035 (+10 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: appending metadata at 1733160419055 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: closing flushed file at 1733160419055Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@799d5592: reopening flushed file at 1733160419068 (+13 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@63367d7e: reopening flushed file at 1733160419075 (+7 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1d2371fb: reopening flushed file at 1733160419081 (+6 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4a8e04b2: reopening flushed file at 1733160419088 (+7 ms)Finished flush of dataSize ~23.25 KB/23805, heapSize ~29.43 KB/30136, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 145ms, sequenceid=60, compaction requested=false at 1733160419095 (+7 ms)Writing region close event to WAL at 1733160419096 (+1 ms)Closed at 1733160419096 2024-12-02T17:26:59,098 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:59,098 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:59,098 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:59,098 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:59,098 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:26:59,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741887_1070 (size=1045) 2024-12-02T17:26:59,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39873 is added to blk_1073741887_1070 (size=1045) 2024-12-02T17:26:59,101 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:26:59,101 INFO [M:0;3b3375d1a925:45913 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T17:26:59,101 INFO [M:0;3b3375d1a925:45913 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:45913 2024-12-02T17:26:59,102 INFO [M:0;3b3375d1a925:45913 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:26:59,172 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741835_1011 (size=393) 2024-12-02T17:26:59,172 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34399 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:26:59,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:59,203 INFO [M:0;3b3375d1a925:45913 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:26:59,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:45913-0x1009c06a5f00000, quorum=127.0.0.1:58139, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:26:59,206 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4d95479b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:59,206 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@469c1f5d{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:59,206 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:59,207 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1191c470{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:59,207 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@13c2f5a4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:59,209 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:59,209 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:59,209 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:59,209 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-797259812-172.17.0.3-1733160376449 (Datanode Uuid 9f811f10-21a0-40f9-b54a-2bca063deae0) service to localhost/127.0.0.1:41131 2024-12-02T17:26:59,209 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data3/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:59,210 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data4/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:59,210 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:59,212 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@f0760d{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:59,213 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@21f536ea{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:59,213 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:59,213 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@495a6aea{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:59,213 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@41559526{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:59,215 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:26:59,215 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:26:59,215 WARN [BP-797259812-172.17.0.3-1733160376449 heartbeating to localhost/127.0.0.1:41131 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-797259812-172.17.0.3-1733160376449 (Datanode Uuid 16a14b74-c77d-4c52-aef6-a3e479cabf82) service to localhost/127.0.0.1:41131 2024-12-02T17:26:59,215 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:26:59,215 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data9/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:59,216 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/cluster_2e912e9c-4955-16c3-db27-01cb1dfad786/data/data10/current/BP-797259812-172.17.0.3-1733160376449 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:26:59,216 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:26:59,223 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5519c514{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:26:59,223 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@47f2ada2{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:26:59,223 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:26:59,223 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@35c95cb4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:26:59,224 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@17c48ca{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir/,STOPPED} 2024-12-02T17:26:59,231 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T17:26:59,261 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T17:26:59,268 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnDatanodeDeath Thread=152 (was 78) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-8-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-18-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:41131 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-21-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:41131 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:41131 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-18-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:35133 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-6-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-5-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-7 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-9-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.1@localhost:35133 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-6-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-8 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-7-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.2@localhost:41131 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:41131 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Timer for 'DataNode' metrics system java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-6-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-11 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.3@localhost:41131 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-10 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:41131 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-21-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-9 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:41131 from jenkins.hfs.2 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-9-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:41131 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-19-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-20-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-19-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-5-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-8-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-20-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-7-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-18-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-5-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:41131 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-20-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-19-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-6 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-7-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-8-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-21-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:41131 from jenkins.hfs.3 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-5 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=448 (was 402) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=220 (was 266), ProcessCount=11 (was 11), AvailableMemoryMB=2300 (was 2319) 2024-12-02T17:26:59,276 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnPipelineRestart Thread=152, OpenFileDescriptor=448, MaxFileDescriptor=1048576, SystemLoadAverage=220, ProcessCount=11, AvailableMemoryMB=2300 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.log.dir so I do NOT create it in target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/46448ab7-7d45-1d4b-cc82-83addfed3535/hadoop.tmp.dir so I do NOT create it in target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4, deleteOnExit=true 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/test.cache.data in system properties and HBase conf 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir in system properties and HBase conf 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T17:26:59,277 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T17:26:59,277 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/nfs.dump.dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/java.io.tmpdir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T17:26:59,278 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T17:26:59,291 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:26:59,374 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:59,380 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:59,383 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:59,383 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:59,383 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:26:59,384 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:59,384 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@affd6eb{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:59,385 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@65e5bfc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:59,498 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7e54696a{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/java.io.tmpdir/jetty-localhost-43291-hadoop-hdfs-3_4_1-tests_jar-_-any-10549849553049884702/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:26:59,499 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@228e200c{HTTP/1.1, (http/1.1)}{localhost:43291} 2024-12-02T17:26:59,499 INFO [Time-limited test {}] server.Server(415): Started @147171ms 2024-12-02T17:26:59,510 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:26:59,578 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:59,582 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:59,583 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:59,583 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:59,583 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:26:59,585 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@371f8296{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:59,585 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5493c38f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:59,708 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2e4cbcc2{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/java.io.tmpdir/jetty-localhost-35109-hadoop-hdfs-3_4_1-tests_jar-_-any-1796509326436605998/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:59,708 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3fab170c{HTTP/1.1, (http/1.1)}{localhost:35109} 2024-12-02T17:26:59,709 INFO [Time-limited test {}] server.Server(415): Started @147381ms 2024-12-02T17:26:59,710 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:59,741 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:26:59,744 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:26:59,746 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:26:59,746 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:26:59,747 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:26:59,747 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@771a779f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:26:59,748 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@50fbcccf{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:26:59,791 WARN [Thread-1172 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data1/current/BP-1459169980-172.17.0.3-1733160419308/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:59,791 WARN [Thread-1173 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data2/current/BP-1459169980-172.17.0.3-1733160419308/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:59,809 WARN [Thread-1151 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:59,812 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x6cb348d33a063adc with lease ID 0xa350cdd2f9ef609a: Processing first storage report for DS-c7b298c2-aca1-40c2-832c-8a55793eeff7 from datanode DatanodeRegistration(127.0.0.1:36027, datanodeUuid=c667f086-f4c3-4afc-ae83-87fcf18d5c75, infoPort=40573, infoSecurePort=0, ipcPort=45787, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308) 2024-12-02T17:26:59,812 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6cb348d33a063adc with lease ID 0xa350cdd2f9ef609a: from storage DS-c7b298c2-aca1-40c2-832c-8a55793eeff7 node DatanodeRegistration(127.0.0.1:36027, datanodeUuid=c667f086-f4c3-4afc-ae83-87fcf18d5c75, infoPort=40573, infoSecurePort=0, ipcPort=45787, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:59,812 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x6cb348d33a063adc with lease ID 0xa350cdd2f9ef609a: Processing first storage report for DS-f89f9348-1e22-4573-b7ee-33e353b60b4e from datanode DatanodeRegistration(127.0.0.1:36027, datanodeUuid=c667f086-f4c3-4afc-ae83-87fcf18d5c75, infoPort=40573, infoSecurePort=0, ipcPort=45787, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308) 2024-12-02T17:26:59,812 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6cb348d33a063adc with lease ID 0xa350cdd2f9ef609a: from storage DS-f89f9348-1e22-4573-b7ee-33e353b60b4e node DatanodeRegistration(127.0.0.1:36027, datanodeUuid=c667f086-f4c3-4afc-ae83-87fcf18d5c75, infoPort=40573, infoSecurePort=0, ipcPort=45787, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:59,865 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7e8ebafe{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/java.io.tmpdir/jetty-localhost-37333-hadoop-hdfs-3_4_1-tests_jar-_-any-2479602352376059123/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:26:59,866 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@2835f29c{HTTP/1.1, (http/1.1)}{localhost:37333} 2024-12-02T17:26:59,866 INFO [Time-limited test {}] server.Server(415): Started @147538ms 2024-12-02T17:26:59,867 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:26:59,942 WARN [Thread-1198 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data3/current/BP-1459169980-172.17.0.3-1733160419308/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:59,943 WARN [Thread-1199 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data4/current/BP-1459169980-172.17.0.3-1733160419308/current, will proceed with Du for space computation calculation, 2024-12-02T17:26:59,963 WARN [Thread-1187 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:26:59,966 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x8d83be1dc3819c5a with lease ID 0xa350cdd2f9ef609b: Processing first storage report for DS-3da1f4b4-7903-4c86-b690-652085799d10 from datanode DatanodeRegistration(127.0.0.1:37343, datanodeUuid=3ab70b2f-12bf-43be-9223-19d2e4e262b2, infoPort=34825, infoSecurePort=0, ipcPort=40491, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308) 2024-12-02T17:26:59,966 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x8d83be1dc3819c5a with lease ID 0xa350cdd2f9ef609b: from storage DS-3da1f4b4-7903-4c86-b690-652085799d10 node DatanodeRegistration(127.0.0.1:37343, datanodeUuid=3ab70b2f-12bf-43be-9223-19d2e4e262b2, infoPort=34825, infoSecurePort=0, ipcPort=40491, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:59,966 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x8d83be1dc3819c5a with lease ID 0xa350cdd2f9ef609b: Processing first storage report for DS-b2249a48-a3e4-431c-9f2d-589b9344e337 from datanode DatanodeRegistration(127.0.0.1:37343, datanodeUuid=3ab70b2f-12bf-43be-9223-19d2e4e262b2, infoPort=34825, infoSecurePort=0, ipcPort=40491, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308) 2024-12-02T17:26:59,966 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x8d83be1dc3819c5a with lease ID 0xa350cdd2f9ef609b: from storage DS-b2249a48-a3e4-431c-9f2d-589b9344e337 node DatanodeRegistration(127.0.0.1:37343, datanodeUuid=3ab70b2f-12bf-43be-9223-19d2e4e262b2, infoPort=34825, infoSecurePort=0, ipcPort=40491, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:26:59,974 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:59,987 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:59,987 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:59,989 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e 2024-12-02T17:26:59,990 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:59,991 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:59,991 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:26:59,992 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/zookeeper_0, clientPort=61144, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T17:26:59,995 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=61144 2024-12-02T17:26:59,996 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:26:59,997 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:00,001 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,002 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,002 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,003 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,010 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:27:00,010 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:27:00,011 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009 with version=8 2024-12-02T17:27:00,011 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase-staging 2024-12-02T17:27:00,014 INFO [Time-limited test {}] client.ConnectionUtils(128): master/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:27:00,014 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:00,014 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:00,014 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:27:00,014 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:00,014 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:27:00,015 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T17:27:00,015 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:27:00,015 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:46397 2024-12-02T17:27:00,017 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:46397 connecting to ZooKeeper ensemble=127.0.0.1:61144 2024-12-02T17:27:00,022 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:463970x0, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:27:00,022 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:46397-0x1009c074d360000 connected 2024-12-02T17:27:00,035 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:00,036 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:00,038 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:27:00,038 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009, hbase.cluster.distributed=false 2024-12-02T17:27:00,039 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:27:00,040 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46397 2024-12-02T17:27:00,040 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46397 2024-12-02T17:27:00,040 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46397 2024-12-02T17:27:00,041 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46397 2024-12-02T17:27:00,041 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46397 2024-12-02T17:27:00,056 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:27:00,056 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:00,056 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:00,056 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:27:00,056 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:00,056 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:27:00,056 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T17:27:00,056 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:27:00,057 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:46459 2024-12-02T17:27:00,058 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:46459 connecting to ZooKeeper ensemble=127.0.0.1:61144 2024-12-02T17:27:00,058 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:00,060 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:00,064 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:464590x0, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:27:00,064 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:46459-0x1009c074d360001 connected 2024-12-02T17:27:00,064 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:27:00,064 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T17:27:00,065 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T17:27:00,066 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T17:27:00,066 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:27:00,067 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46459 2024-12-02T17:27:00,068 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46459 2024-12-02T17:27:00,068 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46459 2024-12-02T17:27:00,068 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46459 2024-12-02T17:27:00,068 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46459 2024-12-02T17:27:00,079 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;3b3375d1a925:46397 2024-12-02T17:27:00,079 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,081 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:27:00,081 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:27:00,081 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,083 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T17:27:00,083 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,083 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,083 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T17:27:00,084 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/3b3375d1a925,46397,1733160420014 from backup master directory 2024-12-02T17:27:00,085 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:27:00,085 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,085 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:27:00,085 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:27:00,085 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,090 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/hbase.id] with ID: 27129de7-b38f-456b-839f-99e8315941f6 2024-12-02T17:27:00,090 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/.tmp/hbase.id 2024-12-02T17:27:00,098 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:27:00,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:27:00,099 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/.tmp/hbase.id]:[hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/hbase.id] 2024-12-02T17:27:00,113 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:00,113 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T17:27:00,115 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-12-02T17:27:00,117 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,117 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:27:00,123 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:27:00,123 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:27:00,124 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T17:27:00,124 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:27:00,134 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:27:00,134 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:27:00,135 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store 2024-12-02T17:27:00,141 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:27:00,141 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:27:00,142 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:00,142 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:27:00,142 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:00,142 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:00,142 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:27:00,142 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:00,142 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:00,142 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160420142Disabling compacts and flushes for region at 1733160420142Disabling writes for close at 1733160420142Writing region close event to WAL at 1733160420142Closed at 1733160420142 2024-12-02T17:27:00,143 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/.initializing 2024-12-02T17:27:00,143 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,145 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C46397%2C1733160420014, suffix=, logDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014, archiveDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/oldWALs, maxLogs=10 2024-12-02T17:27:00,146 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46397%2C1733160420014.1733160420146 2024-12-02T17:27:00,153 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 2024-12-02T17:27:00,159 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34825:34825),(127.0.0.1/127.0.0.1:40573:40573)] 2024-12-02T17:27:00,161 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:27:00,162 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:00,162 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,162 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,164 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,165 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T17:27:00,165 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,166 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,166 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,167 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T17:27:00,167 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,167 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:27:00,168 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,169 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T17:27:00,169 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,169 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:27:00,169 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,170 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T17:27:00,170 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,171 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:27:00,171 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,172 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,172 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,173 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,173 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,174 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T17:27:00,175 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:00,177 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:27:00,178 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=711165, jitterRate=-0.09570786356925964}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T17:27:00,179 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733160420162Initializing all the Stores at 1733160420163 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160420163Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160420164 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160420164Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160420164Cleaning up temporary data from old regions at 1733160420173 (+9 ms)Region opened successfully at 1733160420179 (+6 ms) 2024-12-02T17:27:00,180 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T17:27:00,183 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3eee4295, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:27:00,184 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T17:27:00,184 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T17:27:00,185 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T17:27:00,185 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T17:27:00,185 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T17:27:00,185 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-02T17:27:00,185 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T17:27:00,187 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T17:27:00,188 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T17:27:00,190 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T17:27:00,190 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T17:27:00,191 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T17:27:00,192 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T17:27:00,192 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T17:27:00,194 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T17:27:00,195 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T17:27:00,196 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T17:27:00,198 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T17:27:00,200 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T17:27:00,201 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T17:27:00,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:27:00,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:27:00,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,204 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=3b3375d1a925,46397,1733160420014, sessionid=0x1009c074d360000, setting cluster-up flag (Was=false) 2024-12-02T17:27:00,206 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,207 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,211 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T17:27:00,212 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,216 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,216 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,220 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T17:27:00,221 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,223 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T17:27:00,225 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T17:27:00,225 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T17:27:00,225 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T17:27:00,225 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 3b3375d1a925,46397,1733160420014 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T17:27:00,227 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:27:00,227 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:27:00,227 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:27:00,227 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:27:00,227 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/3b3375d1a925:0, corePoolSize=10, maxPoolSize=10 2024-12-02T17:27:00,227 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,227 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:27:00,227 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,234 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:27:00,234 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T17:27:00,235 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,235 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T17:27:00,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733160450239 2024-12-02T17:27:00,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T17:27:00,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T17:27:00,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T17:27:00,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T17:27:00,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T17:27:00,240 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T17:27:00,240 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,240 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T17:27:00,240 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T17:27:00,240 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T17:27:00,243 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T17:27:00,244 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T17:27:00,244 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160420244,5,FailOnTimeoutGroup] 2024-12-02T17:27:00,244 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160420244,5,FailOnTimeoutGroup] 2024-12-02T17:27:00,244 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,244 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T17:27:00,244 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,245 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,250 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:27:00,250 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:27:00,251 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T17:27:00,251 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009 2024-12-02T17:27:00,260 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:27:00,261 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:27:00,261 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:00,262 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:27:00,264 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:27:00,264 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,264 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,265 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:27:00,266 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:27:00,266 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,266 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,267 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:27:00,268 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:27:00,268 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,268 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,268 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:27:00,270 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:27:00,270 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,270 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,271 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:27:00,271 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(746): ClusterId : 27129de7-b38f-456b-839f-99e8315941f6 2024-12-02T17:27:00,271 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T17:27:00,271 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740 2024-12-02T17:27:00,272 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740 2024-12-02T17:27:00,273 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:27:00,273 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:27:00,273 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T17:27:00,273 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T17:27:00,274 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:27:00,275 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:27:00,277 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T17:27:00,277 DEBUG [RS:0;3b3375d1a925:46459 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1ddd3e74, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:27:00,277 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:27:00,278 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=819499, jitterRate=0.042047545313835144}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:27:00,279 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733160420261Initializing all the Stores at 1733160420262 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160420262Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160420262Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160420262Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160420262Cleaning up temporary data from old regions at 1733160420273 (+11 ms)Region opened successfully at 1733160420279 (+6 ms) 2024-12-02T17:27:00,279 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:27:00,279 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:27:00,279 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:27:00,279 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:27:00,279 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:27:00,279 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:27:00,279 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160420279Disabling compacts and flushes for region at 1733160420279Disabling writes for close at 1733160420279Writing region close event to WAL at 1733160420279Closed at 1733160420279 2024-12-02T17:27:00,281 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:27:00,281 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T17:27:00,281 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T17:27:00,283 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:27:00,284 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T17:27:00,291 DEBUG [RS:0;3b3375d1a925:46459 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;3b3375d1a925:46459 2024-12-02T17:27:00,291 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T17:27:00,291 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T17:27:00,291 DEBUG [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T17:27:00,292 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(2659): reportForDuty to master=3b3375d1a925,46397,1733160420014 with port=46459, startcode=1733160420055 2024-12-02T17:27:00,292 DEBUG [RS:0;3b3375d1a925:46459 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T17:27:00,294 INFO [HMaster-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54925, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T17:27:00,295 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46397 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,295 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46397 {}] master.ServerManager(517): Registering regionserver=3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,297 DEBUG [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009 2024-12-02T17:27:00,297 DEBUG [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:33813 2024-12-02T17:27:00,297 DEBUG [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T17:27:00,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:27:00,299 DEBUG [RS:0;3b3375d1a925:46459 {}] zookeeper.ZKUtil(111): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,299 WARN [RS:0;3b3375d1a925:46459 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:27:00,299 INFO [RS:0;3b3375d1a925:46459 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:27:00,299 DEBUG [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,301 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [3b3375d1a925,46459,1733160420055] 2024-12-02T17:27:00,303 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T17:27:00,306 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T17:27:00,306 INFO [RS:0;3b3375d1a925:46459 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:27:00,306 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,306 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T17:27:00,307 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T17:27:00,307 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,307 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:27:00,308 DEBUG [RS:0;3b3375d1a925:46459 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:27:00,312 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,312 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,312 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,312 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,312 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,312 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46459,1733160420055-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:27:00,335 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T17:27:00,335 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46459,1733160420055-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,335 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,335 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.Replication(171): 3b3375d1a925,46459,1733160420055 started 2024-12-02T17:27:00,357 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,357 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1482): Serving as 3b3375d1a925,46459,1733160420055, RpcServer on 3b3375d1a925/172.17.0.3:46459, sessionid=0x1009c074d360001 2024-12-02T17:27:00,358 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T17:27:00,358 DEBUG [RS:0;3b3375d1a925:46459 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,358 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,46459,1733160420055' 2024-12-02T17:27:00,358 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T17:27:00,358 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T17:27:00,359 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T17:27:00,359 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T17:27:00,359 DEBUG [RS:0;3b3375d1a925:46459 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,359 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,46459,1733160420055' 2024-12-02T17:27:00,359 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T17:27:00,359 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T17:27:00,360 DEBUG [RS:0;3b3375d1a925:46459 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T17:27:00,360 INFO [RS:0;3b3375d1a925:46459 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T17:27:00,360 INFO [RS:0;3b3375d1a925:46459 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T17:27:00,434 WARN [3b3375d1a925:46397 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-02T17:27:00,462 INFO [RS:0;3b3375d1a925:46459 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C46459%2C1733160420055, suffix=, logDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055, archiveDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/oldWALs, maxLogs=32 2024-12-02T17:27:00,463 INFO [RS:0;3b3375d1a925:46459 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46459%2C1733160420055.1733160420462 2024-12-02T17:27:00,469 INFO [RS:0;3b3375d1a925:46459 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 2024-12-02T17:27:00,474 DEBUG [RS:0;3b3375d1a925:46459 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34825:34825),(127.0.0.1/127.0.0.1:40573:40573)] 2024-12-02T17:27:00,504 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:27:00,518 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,519 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,519 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,519 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,519 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,520 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,522 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,522 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,525 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:00,684 DEBUG [3b3375d1a925:46397 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T17:27:00,685 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,686 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,46459,1733160420055, state=OPENING 2024-12-02T17:27:00,688 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T17:27:00,690 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,690 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:00,690 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:27:00,690 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:27:00,690 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,46459,1733160420055}] 2024-12-02T17:27:00,690 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:27:00,843 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T17:27:00,845 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-11-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:58761, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T17:27:00,848 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T17:27:00,848 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:27:00,850 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C46459%2C1733160420055.meta, suffix=.meta, logDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055, archiveDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/oldWALs, maxLogs=32 2024-12-02T17:27:00,850 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta 2024-12-02T17:27:00,855 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta 2024-12-02T17:27:00,856 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34825:34825),(127.0.0.1/127.0.0.1:40573:40573)] 2024-12-02T17:27:00,857 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:27:00,857 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T17:27:00,857 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T17:27:00,857 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T17:27:00,857 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T17:27:00,857 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:00,857 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T17:27:00,857 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T17:27:00,859 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:27:00,859 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:27:00,859 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,860 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,860 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:27:00,861 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:27:00,861 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,861 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,861 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:27:00,862 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:27:00,862 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,863 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,863 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:27:00,863 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:27:00,863 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,864 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:00,864 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:27:00,864 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740 2024-12-02T17:27:00,865 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740 2024-12-02T17:27:00,867 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:27:00,867 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:27:00,867 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:27:00,868 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:27:00,869 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=709776, jitterRate=-0.09747374057769775}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:27:00,869 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T17:27:00,870 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733160420857Writing region info on filesystem at 1733160420857Initializing all the Stores at 1733160420858 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160420858Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160420858Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160420858Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160420858Cleaning up temporary data from old regions at 1733160420867 (+9 ms)Running coprocessor post-open hooks at 1733160420869 (+2 ms)Region opened successfully at 1733160420870 (+1 ms) 2024-12-02T17:27:00,871 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733160420843 2024-12-02T17:27:00,873 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T17:27:00,873 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T17:27:00,874 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,875 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,46459,1733160420055, state=OPEN 2024-12-02T17:27:00,880 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:27:00,880 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:27:00,880 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:27:00,880 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=3b3375d1a925,46459,1733160420055 2024-12-02T17:27:00,880 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:27:00,882 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T17:27:00,883 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,46459,1733160420055 in 190 msec 2024-12-02T17:27:00,885 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T17:27:00,885 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 602 msec 2024-12-02T17:27:00,886 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:27:00,886 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T17:27:00,888 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:27:00,888 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,46459,1733160420055, seqNum=-1] 2024-12-02T17:27:00,888 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:27:00,889 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-11-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49221, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:27:00,894 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 670 msec 2024-12-02T17:27:00,895 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733160420894, completionTime=-1 2024-12-02T17:27:00,895 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T17:27:00,895 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T17:27:00,896 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=1 2024-12-02T17:27:00,896 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733160480896 2024-12-02T17:27:00,896 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733160540896 2024-12-02T17:27:00,896 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 1 msec 2024-12-02T17:27:00,897 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46397,1733160420014-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,897 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46397,1733160420014-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,897 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46397,1733160420014-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,897 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-3b3375d1a925:46397, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,897 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,897 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,898 DEBUG [master/3b3375d1a925:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T17:27:00,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.815sec 2024-12-02T17:27:00,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T17:27:00,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T17:27:00,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T17:27:00,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T17:27:00,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T17:27:00,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46397,1733160420014-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:27:00,901 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46397,1733160420014-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T17:27:00,903 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T17:27:00,903 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T17:27:00,903 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46397,1733160420014-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:00,971 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4c10cfb7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:27:00,971 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 3b3375d1a925,46397,-1 for getting cluster id 2024-12-02T17:27:00,971 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T17:27:00,972 DEBUG [HMaster-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '27129de7-b38f-456b-839f-99e8315941f6' 2024-12-02T17:27:00,973 DEBUG [RPCClient-NioEventLoopGroup-4-12 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T17:27:00,973 DEBUG [RPCClient-NioEventLoopGroup-4-12 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "27129de7-b38f-456b-839f-99e8315941f6" 2024-12-02T17:27:00,973 DEBUG [RPCClient-NioEventLoopGroup-4-12 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5ea85259, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:27:00,973 DEBUG [RPCClient-NioEventLoopGroup-4-12 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [3b3375d1a925,46397,-1] 2024-12-02T17:27:00,973 DEBUG [RPCClient-NioEventLoopGroup-4-12 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T17:27:00,974 DEBUG [RPCClient-NioEventLoopGroup-4-12 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:27:00,975 INFO [HMaster-EventLoopGroup-10-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49228, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T17:27:00,976 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6d866131, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:27:00,976 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:27:00,977 DEBUG [RPCClient-NioEventLoopGroup-4-13 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,46459,1733160420055, seqNum=-1] 2024-12-02T17:27:00,977 DEBUG [RPCClient-NioEventLoopGroup-4-13 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:27:00,978 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-11-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54766, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:27:00,980 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,980 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:00,982 INFO [Time-limited test {}] master.MasterRpcServices(567): Client=null/null set balanceSwitch=false 2024-12-02T17:27:00,982 INFO [Time-limited test {}] wal.TestLogRolling(320): Starting testLogRollOnPipelineRestart 2024-12-02T17:27:00,982 INFO [Time-limited test {}] wal.TestLogRolling(323): Replication=2 2024-12-02T17:27:00,983 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-02T17:27:00,983 DEBUG [RPCClient-NioEventLoopGroup-4-13 {}] client.AsyncConnectionImpl(321): The fetched master address is 3b3375d1a925,46397,1733160420014 2024-12-02T17:27:00,983 DEBUG [RPCClient-NioEventLoopGroup-4-13 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@5b46e379 2024-12-02T17:27:00,983 DEBUG [RPCClient-NioEventLoopGroup-4-13 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T17:27:00,985 INFO [HMaster-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49236, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T17:27:00,985 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46397 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T17:27:00,985 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46397 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T17:27:00,985 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46397 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestLogRolling-testLogRollOnPipelineRestart', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:27:00,987 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46397 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart 2024-12-02T17:27:00,988 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T17:27:00,988 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:00,988 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46397 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRollOnPipelineRestart" procId is: 4 2024-12-02T17:27:00,989 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46397 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:27:00,989 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T17:27:00,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741835_1011 (size=395) 2024-12-02T17:27:00,996 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741835_1011 (size=395) 2024-12-02T17:27:00,997 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 74f1b4c3a01fe1f36cfbe08a071a549e, NAME => 'TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRollOnPipelineRestart', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009 2024-12-02T17:27:01,003 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37343 is added to blk_1073741836_1012 (size=78) 2024-12-02T17:27:01,003 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36027 is added to blk_1073741836_1012 (size=78) 2024-12-02T17:27:01,004 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(898): Instantiated TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:01,004 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1722): Closing 74f1b4c3a01fe1f36cfbe08a071a549e, disabling compactions & flushes 2024-12-02T17:27:01,004 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1755): Closing region TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:01,004 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:01,004 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. after waiting 0 ms 2024-12-02T17:27:01,004 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:01,004 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1973): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:01,004 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1676): Region close journal for 74f1b4c3a01fe1f36cfbe08a071a549e: Waiting for close lock at 1733160421004Disabling compacts and flushes for region at 1733160421004Disabling writes for close at 1733160421004Writing region close event to WAL at 1733160421004Closed at 1733160421004 2024-12-02T17:27:01,005 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T17:27:01,006 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e.","families":{"info":[{"qualifier":"regioninfo","vlen":77,"tag":[],"timestamp":"1733160421005"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733160421005"}]},"ts":"1733160421005"} 2024-12-02T17:27:01,008 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-02T17:27:01,009 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T17:27:01,009 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnPipelineRestart","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160421009"}]},"ts":"1733160421009"} 2024-12-02T17:27:01,011 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testLogRollOnPipelineRestart, state=ENABLING in hbase:meta 2024-12-02T17:27:01,011 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=74f1b4c3a01fe1f36cfbe08a071a549e, ASSIGN}] 2024-12-02T17:27:01,013 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=74f1b4c3a01fe1f36cfbe08a071a549e, ASSIGN 2024-12-02T17:27:01,014 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=74f1b4c3a01fe1f36cfbe08a071a549e, ASSIGN; state=OFFLINE, location=3b3375d1a925,46459,1733160420055; forceNewPlan=false, retain=false 2024-12-02T17:27:01,164 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=74f1b4c3a01fe1f36cfbe08a071a549e, regionState=OPENING, regionLocation=3b3375d1a925,46459,1733160420055 2024-12-02T17:27:01,167 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-11-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=74f1b4c3a01fe1f36cfbe08a071a549e, ASSIGN because future has completed 2024-12-02T17:27:01,167 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 74f1b4c3a01fe1f36cfbe08a071a549e, server=3b3375d1a925,46459,1733160420055}] 2024-12-02T17:27:01,323 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:01,324 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 74f1b4c3a01fe1f36cfbe08a071a549e, NAME => 'TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:27:01,324 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRollOnPipelineRestart 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,324 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:01,324 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,324 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,325 INFO [StoreOpener-74f1b4c3a01fe1f36cfbe08a071a549e-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,327 INFO [StoreOpener-74f1b4c3a01fe1f36cfbe08a071a549e-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 74f1b4c3a01fe1f36cfbe08a071a549e columnFamilyName info 2024-12-02T17:27:01,327 DEBUG [StoreOpener-74f1b4c3a01fe1f36cfbe08a071a549e-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:01,327 INFO [StoreOpener-74f1b4c3a01fe1f36cfbe08a071a549e-1 {}] regionserver.HStore(327): Store=74f1b4c3a01fe1f36cfbe08a071a549e/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:27:01,327 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,328 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,328 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,329 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,329 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,330 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,332 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:27:01,333 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 74f1b4c3a01fe1f36cfbe08a071a549e; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=730350, jitterRate=-0.07131272554397583}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T17:27:01,333 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:01,334 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 74f1b4c3a01fe1f36cfbe08a071a549e: Running coprocessor pre-open hook at 1733160421324Writing region info on filesystem at 1733160421324Initializing all the Stores at 1733160421325 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160421325Cleaning up temporary data from old regions at 1733160421329 (+4 ms)Running coprocessor post-open hooks at 1733160421333 (+4 ms)Region opened successfully at 1733160421334 (+1 ms) 2024-12-02T17:27:01,335 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e., pid=6, masterSystemTime=1733160421320 2024-12-02T17:27:01,337 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:01,337 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:01,338 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=74f1b4c3a01fe1f36cfbe08a071a549e, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,46459,1733160420055 2024-12-02T17:27:01,340 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-11-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 74f1b4c3a01fe1f36cfbe08a071a549e, server=3b3375d1a925,46459,1733160420055 because future has completed 2024-12-02T17:27:01,344 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T17:27:01,344 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 74f1b4c3a01fe1f36cfbe08a071a549e, server=3b3375d1a925,46459,1733160420055 in 174 msec 2024-12-02T17:27:01,346 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T17:27:01,346 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=74f1b4c3a01fe1f36cfbe08a071a549e, ASSIGN in 333 msec 2024-12-02T17:27:01,347 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T17:27:01,347 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnPipelineRestart","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160421347"}]},"ts":"1733160421347"} 2024-12-02T17:27:01,349 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testLogRollOnPipelineRestart, state=ENABLED in hbase:meta 2024-12-02T17:27:01,350 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T17:27:01,352 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart in 364 msec 2024-12-02T17:27:06,360 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:27:06,375 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,375 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,375 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,376 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,376 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,376 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,380 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,380 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,381 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,383 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:06,387 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T17:27:06,388 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRollOnPipelineRestart' 2024-12-02T17:27:07,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:27:07,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-02T17:27:07,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart 2024-12-02T17:27:07,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart Metrics about Tables on a single HBase RegionServer 2024-12-02T17:27:07,868 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:27:07,868 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-02T17:27:11,070 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46397 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:27:11,070 INFO [RPCClient-NioEventLoopGroup-4-15 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestLogRolling-testLogRollOnPipelineRestart completed 2024-12-02T17:27:11,070 DEBUG [Time-limited test {}] hbase.ClientMetaTableAccessor(255): Scanning META starting at row=TestLogRolling-testLogRollOnPipelineRestart,, stopping at row=TestLogRolling-testLogRollOnPipelineRestart ,, for max=2147483647 with caching=100 2024-12-02T17:27:11,073 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2234): Found 1 regions for table TestLogRolling-testLogRollOnPipelineRestart 2024-12-02T17:27:11,073 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2240): firstRegionName=TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:11,077 DEBUG [RPCClient-NioEventLoopGroup-4-14 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestLogRolling-testLogRollOnPipelineRestart', row='row1002', locateType=CURRENT is [region=TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e., hostname=3b3375d1a925,46459,1733160420055, seqNum=2] 2024-12-02T17:27:13,080 INFO [Time-limited test {}] wal.TestLogRolling(360): log.getCurrentFileName()): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 2024-12-02T17:27:13,081 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1009 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1009 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:13,081 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1006 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1006 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:13,081 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1010 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1010 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:13,081 WARN [DataStreamer for file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 block BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1009 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1009 in pipeline [DatanodeInfoWithStorage[127.0.0.1:37343,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK], DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:37343,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]) is bad. 2024-12-02T17:27:13,081 WARN [DataStreamer for file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta block BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1010 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1010 in pipeline [DatanodeInfoWithStorage[127.0.0.1:37343,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK], DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:37343,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]) is bad. 2024-12-02T17:27:13,081 WARN [DataStreamer for file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 block BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1006 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1006 in pipeline [DatanodeInfoWithStorage[127.0.0.1:37343,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK], DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:37343,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]) is bad. 2024-12-02T17:27:13,081 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:37922 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:37343:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37922 dst: /127.0.0.1:37343 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,082 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-839867504_22 at /127.0.0.1:37870 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:37343:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37870 dst: /127.0.0.1:37343 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,082 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:37912 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:37343:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37912 dst: /127.0.0.1:37343 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,082 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:53580 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:36027:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53580 dst: /127.0.0.1:36027 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,082 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-839867504_22 at /127.0.0.1:53546 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:36027:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53546 dst: /127.0.0.1:36027 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,082 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:53582 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:36027:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:53582 dst: /127.0.0.1:36027 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,084 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7e8ebafe{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:13,084 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@2835f29c{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:27:13,085 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:27:13,085 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@50fbcccf{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:27:13,085 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@771a779f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,STOPPED} 2024-12-02T17:27:13,086 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:27:13,086 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:27:13,086 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1459169980-172.17.0.3-1733160419308 (Datanode Uuid 3ab70b2f-12bf-43be-9223-19d2e4e262b2) service to localhost/127.0.0.1:33813 2024-12-02T17:27:13,086 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:27:13,086 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data3/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:13,087 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data4/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:13,087 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:27:13,104 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:27:13,109 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:27:13,109 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:27:13,109 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:27:13,109 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:27:13,110 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@12f241e2{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:27:13,110 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@55cf8f7d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:27:13,223 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3bf6b78a{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/java.io.tmpdir/jetty-localhost-35403-hadoop-hdfs-3_4_1-tests_jar-_-any-700633257316694198/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:13,223 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1dd2ccd0{HTTP/1.1, (http/1.1)}{localhost:35403} 2024-12-02T17:27:13,223 INFO [Time-limited test {}] server.Server(415): Started @160896ms 2024-12-02T17:27:13,225 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:27:13,245 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1013 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1013 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:13,245 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1014 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1014 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:13,245 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1015 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1015 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:13,246 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-839867504_22 at /127.0.0.1:38912 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:36027:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:38912 dst: /127.0.0.1:36027 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,247 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:38910 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:36027:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:38910 dst: /127.0.0.1:36027 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,247 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:38904 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:36027:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:38904 dst: /127.0.0.1:36027 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:13,255 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2e4cbcc2{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:13,255 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3fab170c{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:27:13,255 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:27:13,255 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5493c38f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:27:13,255 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@371f8296{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,STOPPED} 2024-12-02T17:27:13,257 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:27:13,257 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:27:13,257 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1459169980-172.17.0.3-1733160419308 (Datanode Uuid c667f086-f4c3-4afc-ae83-87fcf18d5c75) service to localhost/127.0.0.1:33813 2024-12-02T17:27:13,257 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:27:13,257 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data1/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:13,257 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data2/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:13,258 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:27:13,276 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:27:13,281 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:27:13,282 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:27:13,282 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:27:13,282 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:27:13,283 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3f4861a6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:27:13,283 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@e87d64e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:27:13,322 WARN [Thread-1323 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:27:13,324 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x15d9739ed39157fc with lease ID 0xa350cdd2f9ef609c: from storage DS-3da1f4b4-7903-4c86-b690-652085799d10 node DatanodeRegistration(127.0.0.1:45887, datanodeUuid=3ab70b2f-12bf-43be-9223-19d2e4e262b2, infoPort=44499, infoSecurePort=0, ipcPort=37653, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 6, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:13,324 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x15d9739ed39157fc with lease ID 0xa350cdd2f9ef609c: from storage DS-b2249a48-a3e4-431c-9f2d-589b9344e337 node DatanodeRegistration(127.0.0.1:45887, datanodeUuid=3ab70b2f-12bf-43be-9223-19d2e4e262b2, infoPort=44499, infoSecurePort=0, ipcPort=37653, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 6, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:13,395 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@355539f9{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/java.io.tmpdir/jetty-localhost-37971-hadoop-hdfs-3_4_1-tests_jar-_-any-9071207856576006256/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:13,396 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3d3d6246{HTTP/1.1, (http/1.1)}{localhost:37971} 2024-12-02T17:27:13,396 INFO [Time-limited test {}] server.Server(415): Started @161068ms 2024-12-02T17:27:13,397 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:27:13,472 WARN [Thread-1354 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:27:13,474 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xf3be410bf7c7b396 with lease ID 0xa350cdd2f9ef609d: from storage DS-c7b298c2-aca1-40c2-832c-8a55793eeff7 node DatanodeRegistration(127.0.0.1:40691, datanodeUuid=c667f086-f4c3-4afc-ae83-87fcf18d5c75, infoPort=38121, infoSecurePort=0, ipcPort=38929, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 6, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:13,474 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xf3be410bf7c7b396 with lease ID 0xa350cdd2f9ef609d: from storage DS-f89f9348-1e22-4573-b7ee-33e353b60b4e node DatanodeRegistration(127.0.0.1:40691, datanodeUuid=c667f086-f4c3-4afc-ae83-87fcf18d5c75, infoPort=38121, infoSecurePort=0, ipcPort=38929, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 6, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:14,414 INFO [Time-limited test {}] wal.TestLogRolling(372): Data Nodes restarted 2024-12-02T17:27:14,416 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1002 2024-12-02T17:27:14,418 ERROR [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009-prefix:3b3375d1a925,46459,1733160420055 {}] wal.AbstractFSWAL(1838): appendAndSync throws IOException. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:14,418 WARN [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009-prefix:3b3375d1a925,46459,1733160420055 {}] wal.AbstractFSWAL(2174): append entry failed java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:14,418 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C46459%2C1733160420055:(num 1733160420462) roll requested 2024-12-02T17:27:14,418 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46459%2C1733160420055.1733160434418 2024-12-02T17:27:14,424 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.TestLogRolling$2(347): preLogRoll: oldFile=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 newFile=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 2024-12-02T17:27:14,424 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:14,424 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:14,424 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:14,424 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:14,424 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:14,425 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 with entries=2, filesize=1.59 KB; new WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 2024-12-02T17:27:14,425 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:14,425 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:14,425 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 2024-12-02T17:27:14,425 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38121:38121),(127.0.0.1/127.0.0.1:44499:44499)] 2024-12-02T17:27:14,425 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 is not closed yet, will try archiving it next time 2024-12-02T17:27:14,425 WARN [IPC Server handler 3 on default port 33813 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 has not been closed. Lease recovery is in progress. RecoveryId = 1017 for block blk_1073741833_1013 2024-12-02T17:27:14,426 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 after 1ms 2024-12-02T17:27:15,325 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741833_1013: GenerationStamp not matched, existing replica is blk_1073741833_1009 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-02T17:27:16,431 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1003 2024-12-02T17:27:18,426 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 after 4001ms 2024-12-02T17:27:18,434 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1016 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1016 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:18,434 WARN [DataStreamer for file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 block BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1016 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1016 in pipeline [DatanodeInfoWithStorage[127.0.0.1:40691,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK], DatanodeInfoWithStorage[127.0.0.1:45887,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:40691,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]) is bad. 2024-12-02T17:27:18,435 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:36664 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1016] {}] datanode.DataXceiver(331): 127.0.0.1:40691:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36664 dst: /127.0.0.1:40691 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:18,435 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:38948 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1016] {}] datanode.DataXceiver(331): 127.0.0.1:45887:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:38948 dst: /127.0.0.1:45887 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:18,436 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@355539f9{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:18,436 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3d3d6246{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:27:18,437 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:27:18,437 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@e87d64e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:27:18,437 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3f4861a6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,STOPPED} 2024-12-02T17:27:18,438 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:27:18,438 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:27:18,438 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:27:18,438 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1459169980-172.17.0.3-1733160419308 (Datanode Uuid c667f086-f4c3-4afc-ae83-87fcf18d5c75) service to localhost/127.0.0.1:33813 2024-12-02T17:27:18,439 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data1/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:18,439 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data2/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:18,439 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:27:18,450 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:27:18,453 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:27:18,454 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:27:18,454 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:27:18,454 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:27:18,455 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@67d70e61{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:27:18,455 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@290ac13e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:27:18,576 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5036ecf5{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/java.io.tmpdir/jetty-localhost-40783-hadoop-hdfs-3_4_1-tests_jar-_-any-14009281473140296584/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:18,576 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@2647e756{HTTP/1.1, (http/1.1)}{localhost:40783} 2024-12-02T17:27:18,576 INFO [Time-limited test {}] server.Server(415): Started @166249ms 2024-12-02T17:27:18,577 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:27:18,601 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1018 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1018 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:18,601 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1410848226_22 at /127.0.0.1:50398 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741837_1016] {}] datanode.DataXceiver(331): 127.0.0.1:45887:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50398 dst: /127.0.0.1:45887 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:18,603 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3bf6b78a{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:18,604 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1dd2ccd0{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:27:18,604 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:27:18,604 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@55cf8f7d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:27:18,604 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@12f241e2{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,STOPPED} 2024-12-02T17:27:18,605 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:27:18,605 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:27:18,605 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1459169980-172.17.0.3-1733160419308 (Datanode Uuid 3ab70b2f-12bf-43be-9223-19d2e4e262b2) service to localhost/127.0.0.1:33813 2024-12-02T17:27:18,605 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:27:18,606 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data3/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:18,606 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data4/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:18,606 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:27:18,614 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:27:18,617 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:27:18,618 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:27:18,618 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:27:18,618 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:27:18,618 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@652ca842{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:27:18,619 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3172a7c6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:27:18,710 WARN [Thread-1397 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:27:18,712 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9464e99ec9200647 with lease ID 0xa350cdd2f9ef609e: from storage DS-c7b298c2-aca1-40c2-832c-8a55793eeff7 node DatanodeRegistration(127.0.0.1:37039, datanodeUuid=c667f086-f4c3-4afc-ae83-87fcf18d5c75, infoPort=44329, infoSecurePort=0, ipcPort=46099, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:18,712 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9464e99ec9200647 with lease ID 0xa350cdd2f9ef609e: from storage DS-f89f9348-1e22-4573-b7ee-33e353b60b4e node DatanodeRegistration(127.0.0.1:37039, datanodeUuid=c667f086-f4c3-4afc-ae83-87fcf18d5c75, infoPort=44329, infoSecurePort=0, ipcPort=46099, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 6, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:18,769 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1ed6e4b8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/java.io.tmpdir/jetty-localhost-46629-hadoop-hdfs-3_4_1-tests_jar-_-any-16804985910934554565/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:18,769 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@929fb8a{HTTP/1.1, (http/1.1)}{localhost:46629} 2024-12-02T17:27:18,769 INFO [Time-limited test {}] server.Server(415): Started @166441ms 2024-12-02T17:27:18,771 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:27:18,858 WARN [Thread-1428 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:27:18,860 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9aed7af6d0ebf6ff with lease ID 0xa350cdd2f9ef609f: from storage DS-3da1f4b4-7903-4c86-b690-652085799d10 node DatanodeRegistration(127.0.0.1:42681, datanodeUuid=3ab70b2f-12bf-43be-9223-19d2e4e262b2, infoPort=36379, infoSecurePort=0, ipcPort=35935, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 6, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:18,861 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9aed7af6d0ebf6ff with lease ID 0xa350cdd2f9ef609f: from storage DS-b2249a48-a3e4-431c-9f2d-589b9344e337 node DatanodeRegistration(127.0.0.1:42681, datanodeUuid=3ab70b2f-12bf-43be-9223-19d2e4e262b2, infoPort=36379, infoSecurePort=0, ipcPort=35935, storageInfo=lv=-57;cid=testClusterID;nsid=858190954;c=1733160419308), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:19,787 INFO [Time-limited test {}] wal.TestLogRolling(389): Data Nodes restarted 2024-12-02T17:27:19,789 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1004 2024-12-02T17:27:19,791 ERROR [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009-prefix:3b3375d1a925,46459,1733160420055 {}] wal.AbstractFSWAL(1838): appendAndSync throws IOException. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:45887,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:19,791 WARN [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009-prefix:3b3375d1a925,46459,1733160420055 {}] wal.AbstractFSWAL(2174): append entry failed java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:45887,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:19,791 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C46459%2C1733160420055:(num 1733160434418) roll requested 2024-12-02T17:27:19,791 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46459%2C1733160420055.1733160439791 2024-12-02T17:27:19,802 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.TestLogRolling$2(347): preLogRoll: oldFile=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 newFile=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 2024-12-02T17:27:19,802 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:19,803 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:19,803 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:19,803 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:19,803 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:19,803 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 with entries=2, filesize=2.37 KB; new WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 2024-12-02T17:27:19,803 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:45887,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:19,803 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:45887,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:19,804 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 2024-12-02T17:27:19,804 WARN [IPC Server handler 3 on default port 33813 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 has not been closed. Lease recovery is in progress. RecoveryId = 1020 for block blk_1073741837_1018 2024-12-02T17:27:19,804 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 after 0ms 2024-12-02T17:27:19,804 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:44329:44329),(127.0.0.1/127.0.0.1:36379:36379)] 2024-12-02T17:27:19,804 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 is not closed yet, will try archiving it next time 2024-12-02T17:27:19,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741837_1020 (size=2427) 2024-12-02T17:27:21,806 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:21,812 DEBUG [Time-limited test {}] wal.TestLogRolling$2(347): preLogRoll: oldFile=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 newFile=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:21,812 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:21,812 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:21,812 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:21,812 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:21,812 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:21,812 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 with entries=1, filesize=1.23 KB; new WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:21,813 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:36379:36379),(127.0.0.1/127.0.0.1:44329:44329)] 2024-12-02T17:27:21,813 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 is not closed yet, will try archiving it next time 2024-12-02T17:27:21,813 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 is not closed yet, will try archiving it next time 2024-12-02T17:27:21,814 DEBUG [Time-limited test {}] wal.TestLogRolling(403): recovering lease for hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 2024-12-02T17:27:21,814 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 2024-12-02T17:27:21,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741838_1019 (size=1264) 2024-12-02T17:27:21,814 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 after 0ms 2024-12-02T17:27:21,814 DEBUG [Time-limited test {}] wal.TestLogRolling(407): Reading WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 2024-12-02T17:27:21,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741838_1019 (size=1264) 2024-12-02T17:27:21,815 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(879): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 is not closed yet, will try archiving it next time 2024-12-02T17:27:21,823 DEBUG [Time-limited test {}] wal.TestLogRolling(412): #3: [\x00/METAFAMILY:HBASE::REGION_EVENT::REGION_OPEN/1733160421334/Put/vlen=218/seqid=0] 2024-12-02T17:27:21,824 DEBUG [Time-limited test {}] wal.TestLogRolling(412): #4: [row1002/info:/1733160431078/Put/vlen=1045/seqid=0] 2024-12-02T17:27:21,824 DEBUG [Time-limited test {}] wal.TestLogRolling(419): EOF reading file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160420462 2024-12-02T17:27:21,824 DEBUG [Time-limited test {}] wal.TestLogRolling(403): recovering lease for hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 2024-12-02T17:27:21,824 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 2024-12-02T17:27:21,824 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 after 0ms 2024-12-02T17:27:21,824 DEBUG [Time-limited test {}] wal.TestLogRolling(407): Reading WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 2024-12-02T17:27:21,827 DEBUG [Time-limited test {}] wal.TestLogRolling(412): #5: [row1003/info:/1733160434417/Put/vlen=1045/seqid=0] 2024-12-02T17:27:21,827 DEBUG [Time-limited test {}] wal.TestLogRolling(412): #6: [row1004/info:/1733160436432/Put/vlen=1045/seqid=0] 2024-12-02T17:27:21,827 DEBUG [Time-limited test {}] wal.TestLogRolling(419): EOF reading file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 2024-12-02T17:27:21,827 DEBUG [Time-limited test {}] wal.TestLogRolling(403): recovering lease for hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 2024-12-02T17:27:21,827 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 2024-12-02T17:27:21,828 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 after 1ms 2024-12-02T17:27:21,828 DEBUG [Time-limited test {}] wal.TestLogRolling(407): Reading WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160439791 2024-12-02T17:27:21,830 DEBUG [Time-limited test {}] wal.TestLogRolling(412): #7: [row1005/info:/1733160439790/Put/vlen=1045/seqid=0] 2024-12-02T17:27:21,830 DEBUG [Time-limited test {}] wal.TestLogRolling(403): recovering lease for hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:21,830 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:21,831 WARN [IPC Server handler 0 on default port 33813 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 has not been closed. Lease recovery is in progress. RecoveryId = 1022 for block blk_1073741839_1021 2024-12-02T17:27:21,831 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 after 0ms 2024-12-02T17:27:22,718 WARN [ResponseProcessor for block BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:22,718 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-839867504_22 at /127.0.0.1:57422 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021] {}] datanode.DataXceiver(331): 127.0.0.1:42681:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:57422 dst: /127.0.0.1:42681 java.io.InterruptedIOException: Interrupted while waiting for IO on channel java.nio.channels.SocketChannel[connected local=localhost/127.0.0.1:42681 remote=/127.0.0.1:57422]. Total timeout mills is 60000, 59093 millis timeout left. at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:350) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:22,718 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-839867504_22 at /127.0.0.1:41826 [Receiving block BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021] {}] datanode.DataXceiver(331): 127.0.0.1:37039:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41826 dst: /127.0.0.1:37039 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T17:27:22,718 WARN [DataStreamer for file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 block BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 in pipeline [DatanodeInfoWithStorage[127.0.0.1:42681,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK], DatanodeInfoWithStorage[127.0.0.1:37039,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:42681,DS-3da1f4b4-7903-4c86-b690-652085799d10,DISK]) is bad. 2024-12-02T17:27:22,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741839_1022 (size=85) 2024-12-02T17:27:22,723 WARN [DataStreamer for file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 block BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 {}] hdfs.DataStreamer(859): DataStreamer Exception org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:23,716 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741837_1018: GenerationStamp not matched, existing replica is blk_1073741837_1016 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-02T17:27:23,805 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160434418 after 4001ms 2024-12-02T17:27:25,832 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 after 4001ms 2024-12-02T17:27:25,832 DEBUG [Time-limited test {}] wal.TestLogRolling(407): Reading WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:25,836 DEBUG [Time-limited test {}] wal.TestLogRolling(419): EOF reading file /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:25,836 INFO [Time-limited test {}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.74 KB heapSize=3.77 KB 2024-12-02T17:27:25,836 ERROR [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009-prefix:3b3375d1a925,46459,1733160420055.meta {}] wal.AbstractFSWAL(1838): appendAndSync throws IOException. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:25,837 WARN [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009-prefix:3b3375d1a925,46459,1733160420055.meta {}] wal.AbstractFSWAL(2174): append entry failed java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:25,837 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C46459%2C1733160420055.meta:.meta(num 1733160420850) roll requested 2024-12-02T17:27:25,837 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46459%2C1733160420055.meta.1733160445837.meta 2024-12-02T17:27:25,845 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,846 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,846 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,846 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,846 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,846 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta with entries=8, filesize=2.36 KB; new WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160445837.meta 2024-12-02T17:27:25,846 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:25,846 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:25,847 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta 2024-12-02T17:27:25,847 WARN [IPC Server handler 2 on default port 33813 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta has not been closed. Lease recovery is in progress. RecoveryId = 1024 for block blk_1073741834_1014 2024-12-02T17:27:25,847 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta after 0ms 2024-12-02T17:27:25,847 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:36379:36379),(127.0.0.1/127.0.0.1:44329:44329)] 2024-12-02T17:27:25,847 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta is not closed yet, will try archiving it next time 2024-12-02T17:27:25,866 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/info/a475d9fd8c04410791763cf4718caea2 is 207, key is TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e./info:regioninfo/1733160421337/Put/seqid=0 2024-12-02T17:27:25,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741841_1025 (size=7125) 2024-12-02T17:27:25,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741841_1025 (size=7125) 2024-12-02T17:27:25,871 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.52 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/info/a475d9fd8c04410791763cf4718caea2 2024-12-02T17:27:25,890 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/ns/6150a150ad754614ab85afab0859d8a5 is 43, key is default/ns:d/1733160420890/Put/seqid=0 2024-12-02T17:27:25,894 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741842_1026 (size=5153) 2024-12-02T17:27:25,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741842_1026 (size=5153) 2024-12-02T17:27:25,895 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/ns/6150a150ad754614ab85afab0859d8a5 2024-12-02T17:27:25,914 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/table/cd7fc45fc9a843ad8a41c5a475980ef3 is 79, key is TestLogRolling-testLogRollOnPipelineRestart/table:state/1733160421347/Put/seqid=0 2024-12-02T17:27:25,918 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741843_1027 (size=5438) 2024-12-02T17:27:25,918 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741843_1027 (size=5438) 2024-12-02T17:27:25,919 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=150 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/table/cd7fc45fc9a843ad8a41c5a475980ef3 2024-12-02T17:27:25,924 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/info/a475d9fd8c04410791763cf4718caea2 as hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/info/a475d9fd8c04410791763cf4718caea2 2024-12-02T17:27:25,930 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/info/a475d9fd8c04410791763cf4718caea2, entries=10, sequenceid=11, filesize=7.0 K 2024-12-02T17:27:25,931 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/ns/6150a150ad754614ab85afab0859d8a5 as hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/ns/6150a150ad754614ab85afab0859d8a5 2024-12-02T17:27:25,936 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/ns/6150a150ad754614ab85afab0859d8a5, entries=2, sequenceid=11, filesize=5.0 K 2024-12-02T17:27:25,937 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/.tmp/table/cd7fc45fc9a843ad8a41c5a475980ef3 as hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/table/cd7fc45fc9a843ad8a41c5a475980ef3 2024-12-02T17:27:25,942 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/table/cd7fc45fc9a843ad8a41c5a475980ef3, entries=2, sequenceid=11, filesize=5.3 K 2024-12-02T17:27:25,943 INFO [Time-limited test {}] regionserver.HRegion(3140): Finished flush of dataSize ~1.74 KB/1782, heapSize ~3.48 KB/3560, currentSize=0 B/0 for 1588230740 in 107ms, sequenceid=11, compaction requested=false 2024-12-02T17:27:25,943 DEBUG [Time-limited test {}] regionserver.HRegion(2603): Flush status journal for 1588230740: 2024-12-02T17:27:25,943 INFO [Time-limited test {}] regionserver.HRegion(2902): Flushing 74f1b4c3a01fe1f36cfbe08a071a549e 1/1 column families, dataSize=4.20 KB heapSize=4.75 KB 2024-12-02T17:27:25,943 ERROR [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009-prefix:3b3375d1a925,46459,1733160420055 {}] wal.AbstractFSWAL(1838): appendAndSync throws IOException. org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:25,944 WARN [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009-prefix:3b3375d1a925,46459,1733160420055 {}] wal.AbstractFSWAL(2174): append entry failed org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:25,944 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C46459%2C1733160420055:(num 1733160441806) roll requested 2024-12-02T17:27:25,944 INFO [regionserver/3b3375d1a925:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46459%2C1733160420055.1733160445944 2024-12-02T17:27:25,949 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.TestLogRolling$2(347): preLogRoll: oldFile=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 newFile=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160445944 2024-12-02T17:27:25,949 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,949 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,949 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,949 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,950 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:25,950 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160445944 2024-12-02T17:27:25,950 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:25,950 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1459169980-172.17.0.3-1733160419308:blk_1073741839_1021 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor102.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:25,951 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:36379:36379),(127.0.0.1/127.0.0.1:44329:44329)] 2024-12-02T17:27:25,951 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:25,951 DEBUG [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractFSWAL(879): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 is not closed yet, will try archiving it next time 2024-12-02T17:27:25,951 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 after 0ms 2024-12-02T17:27:25,951 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.1733160441806 to hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/oldWALs/3b3375d1a925%2C46459%2C1733160420055.1733160441806 2024-12-02T17:27:25,965 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e/.tmp/info/e766478893a5471d9c4868b01c52b30c is 1080, key is row1002/info:/1733160431078/Put/seqid=0 2024-12-02T17:27:25,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741845_1029 (size=9270) 2024-12-02T17:27:25,971 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=4.20 KB at sequenceid=8 (bloomFilter=true), to=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e/.tmp/info/e766478893a5471d9c4868b01c52b30c 2024-12-02T17:27:25,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741845_1029 (size=9270) 2024-12-02T17:27:25,977 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e/.tmp/info/e766478893a5471d9c4868b01c52b30c as hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e/info/e766478893a5471d9c4868b01c52b30c 2024-12-02T17:27:25,981 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e/info/e766478893a5471d9c4868b01c52b30c, entries=4, sequenceid=8, filesize=9.1 K 2024-12-02T17:27:25,983 INFO [Time-limited test {}] regionserver.HRegion(3140): Finished flush of dataSize ~4.20 KB/4304, heapSize ~4.73 KB/4848, currentSize=0 B/0 for 74f1b4c3a01fe1f36cfbe08a071a549e in 39ms, sequenceid=8, compaction requested=false 2024-12-02T17:27:25,983 DEBUG [Time-limited test {}] regionserver.HRegion(2603): Flush status journal for 74f1b4c3a01fe1f36cfbe08a071a549e: 2024-12-02T17:27:25,987 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T17:27:25,987 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:27:25,987 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:27:25,987 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:27:25,988 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:27:25,988 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T17:27:25,988 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T17:27:25,988 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=2056979492, stopped=false 2024-12-02T17:27:25,988 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=3b3375d1a925,46397,1733160420014 2024-12-02T17:27:25,990 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:27:25,990 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:27:25,990 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:25,990 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:25,990 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:27:25,990 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:27:25,990 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:27:25,990 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:27:25,991 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:27:25,991 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:27:25,991 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '3b3375d1a925,46459,1733160420055' ***** 2024-12-02T17:27:25,991 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T17:27:25,991 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T17:27:25,991 INFO [RS:0;3b3375d1a925:46459 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T17:27:25,991 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T17:27:25,991 INFO [RS:0;3b3375d1a925:46459 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T17:27:25,991 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(3091): Received CLOSE for 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:25,992 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(959): stopping server 3b3375d1a925,46459,1733160420055 2024-12-02T17:27:25,992 INFO [RS:0;3b3375d1a925:46459 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:27:25,992 INFO [RS:0;3b3375d1a925:46459 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;3b3375d1a925:46459. 2024-12-02T17:27:25,992 DEBUG [RS:0;3b3375d1a925:46459 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:27:25,992 DEBUG [RS:0;3b3375d1a925:46459 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:27:25,992 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 74f1b4c3a01fe1f36cfbe08a071a549e, disabling compactions & flushes 2024-12-02T17:27:25,992 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:25,992 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T17:27:25,992 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:25,992 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T17:27:25,992 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T17:27:25,992 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. after waiting 0 ms 2024-12-02T17:27:25,992 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:25,992 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T17:27:25,992 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-02T17:27:25,992 DEBUG [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740, 74f1b4c3a01fe1f36cfbe08a071a549e=TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e.} 2024-12-02T17:27:25,992 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:27:25,993 DEBUG [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, 74f1b4c3a01fe1f36cfbe08a071a549e 2024-12-02T17:27:25,993 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:27:25,993 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:27:25,993 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:27:25,993 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:27:25,997 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/default/TestLogRolling-testLogRollOnPipelineRestart/74f1b4c3a01fe1f36cfbe08a071a549e/recovered.edits/11.seqid, newMaxSeqId=11, maxSeqId=1 2024-12-02T17:27:25,997 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:25,997 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 74f1b4c3a01fe1f36cfbe08a071a549e: Waiting for close lock at 1733160445992Running coprocessor pre-close hooks at 1733160445992Disabling compacts and flushes for region at 1733160445992Disabling writes for close at 1733160445992Writing region close event to WAL at 1733160445993 (+1 ms)Running coprocessor post-close hooks at 1733160445997 (+4 ms)Closed at 1733160445997 2024-12-02T17:27:25,998 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733160420985.74f1b4c3a01fe1f36cfbe08a071a549e. 2024-12-02T17:27:25,998 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-02T17:27:25,998 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:27:25,998 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:27:25,998 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160445992Running coprocessor pre-close hooks at 1733160445992Disabling compacts and flushes for region at 1733160445992Disabling writes for close at 1733160445993 (+1 ms)Writing region close event to WAL at 1733160445994 (+1 ms)Running coprocessor post-close hooks at 1733160445998 (+4 ms)Closed at 1733160445998 2024-12-02T17:27:25,999 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T17:27:26,193 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(976): stopping server 3b3375d1a925,46459,1733160420055; all regions closed. 2024-12-02T17:27:26,193 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:26,193 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:26,193 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:26,193 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:26,194 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:26,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741840_1023 (size=825) 2024-12-02T17:27:26,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741840_1023 (size=825) 2024-12-02T17:27:26,313 INFO [regionserver/3b3375d1a925:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:27:26,326 INFO [regionserver/3b3375d1a925:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-02T17:27:26,326 INFO [regionserver/3b3375d1a925:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-02T17:27:27,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:27:27,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:27:27,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart 2024-12-02T17:27:29,848 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta after 4001ms 2024-12-02T17:27:29,848 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/WALs/3b3375d1a925,46459,1733160420055/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta to hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/oldWALs/3b3375d1a925%2C46459%2C1733160420055.meta.1733160420850.meta 2024-12-02T17:27:29,851 DEBUG [RS:0;3b3375d1a925:46459 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/oldWALs 2024-12-02T17:27:29,851 INFO [RS:0;3b3375d1a925:46459 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C46459%2C1733160420055.meta:.meta(num 1733160445837) 2024-12-02T17:27:29,851 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,851 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,852 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,852 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,852 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,853 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741844_1028 (size=1162) 2024-12-02T17:27:29,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741844_1028 (size=1162) 2024-12-02T17:27:29,858 DEBUG [RS:0;3b3375d1a925:46459 {}] wal.AbstractFSWAL(1256): Moved 4 WAL file(s) to /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/oldWALs 2024-12-02T17:27:29,858 INFO [RS:0;3b3375d1a925:46459 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C46459%2C1733160420055:(num 1733160445944) 2024-12-02T17:27:29,858 DEBUG [RS:0;3b3375d1a925:46459 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:27:29,858 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:27:29,858 INFO [RS:0;3b3375d1a925:46459 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:27:29,858 INFO [RS:0;3b3375d1a925:46459 {}] hbase.ChoreService(370): Chore service for: regionserver/3b3375d1a925:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-02T17:27:29,859 INFO [RS:0;3b3375d1a925:46459 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:27:29,859 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:27:29,859 INFO [RS:0;3b3375d1a925:46459 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:46459 2024-12-02T17:27:29,860 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741834_1014: GenerationStamp not matched, existing replica is blk_1073741834_1010 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-02T17:27:29,861 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/3b3375d1a925,46459,1733160420055 2024-12-02T17:27:29,861 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:27:29,861 INFO [RS:0;3b3375d1a925:46459 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:27:29,862 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [3b3375d1a925,46459,1733160420055] 2024-12-02T17:27:29,864 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/3b3375d1a925,46459,1733160420055 already deleted, retry=false 2024-12-02T17:27:29,864 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 3b3375d1a925,46459,1733160420055 expired; onlineServers=0 2024-12-02T17:27:29,864 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '3b3375d1a925,46397,1733160420014' ***** 2024-12-02T17:27:29,864 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T17:27:29,864 INFO [M:0;3b3375d1a925:46397 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:27:29,864 INFO [M:0;3b3375d1a925:46397 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:27:29,864 DEBUG [M:0;3b3375d1a925:46397 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T17:27:29,864 DEBUG [M:0;3b3375d1a925:46397 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T17:27:29,864 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T17:27:29,864 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160420244 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160420244,5,FailOnTimeoutGroup] 2024-12-02T17:27:29,864 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160420244 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160420244,5,FailOnTimeoutGroup] 2024-12-02T17:27:29,864 INFO [M:0;3b3375d1a925:46397 {}] hbase.ChoreService(370): Chore service for: master/3b3375d1a925:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T17:27:29,864 INFO [M:0;3b3375d1a925:46397 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:27:29,865 DEBUG [M:0;3b3375d1a925:46397 {}] master.HMaster(1795): Stopping service threads 2024-12-02T17:27:29,865 INFO [M:0;3b3375d1a925:46397 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T17:27:29,865 INFO [M:0;3b3375d1a925:46397 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:27:29,865 INFO [M:0;3b3375d1a925:46397 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T17:27:29,865 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T17:27:29,865 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T17:27:29,865 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:29,865 DEBUG [M:0;3b3375d1a925:46397 {}] zookeeper.ZKUtil(347): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T17:27:29,865 WARN [M:0;3b3375d1a925:46397 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T17:27:29,866 INFO [M:0;3b3375d1a925:46397 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/.lastflushedseqids 2024-12-02T17:27:29,875 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741846_1030 (size=139) 2024-12-02T17:27:29,875 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741846_1030 (size=139) 2024-12-02T17:27:29,876 INFO [M:0;3b3375d1a925:46397 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T17:27:29,876 INFO [M:0;3b3375d1a925:46397 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T17:27:29,876 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:27:29,876 INFO [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:29,876 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:29,876 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:27:29,876 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:29,876 INFO [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=23.17 KB heapSize=29.16 KB 2024-12-02T17:27:29,877 ERROR [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData-prefix:3b3375d1a925,46397,1733160420014 {}] wal.AbstractFSWAL(1838): appendAndSync throws IOException. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:29,877 WARN [FSHLog-0-hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData-prefix:3b3375d1a925,46397,1733160420014 {}] wal.AbstractFSWAL(2174): append entry failed java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:29,877 DEBUG [master:store-WAL-Roller {}] wal.AbstractWALRoller(198): WAL FSHLog 3b3375d1a925%2C46397%2C1733160420014:(num 1733160420146) roll requested 2024-12-02T17:27:29,877 INFO [master:store-WAL-Roller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46397%2C1733160420014.1733160449877 2024-12-02T17:27:29,882 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,882 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,883 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,883 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,883 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:29,883 INFO [master:store-WAL-Roller {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 with entries=53, filesize=26.62 KB; new WAL /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160449877 2024-12-02T17:27:29,884 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:29,885 WARN [Close-WAL-Writer-0 {}] wal.AbstractFSWAL(2043): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36027,DS-c7b298c2-aca1-40c2-832c-8a55793eeff7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T17:27:29,885 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 2024-12-02T17:27:29,885 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:44329:44329),(127.0.0.1/127.0.0.1:36379:36379)] 2024-12-02T17:27:29,885 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(879): hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 is not closed yet, will try archiving it next time 2024-12-02T17:27:29,885 WARN [IPC Server handler 3 on default port 33813 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 has not been closed. Lease recovery is in progress. RecoveryId = 1032 for block blk_1073741830_1015 2024-12-02T17:27:29,885 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 after 0ms 2024-12-02T17:27:29,900 DEBUG [M:0;3b3375d1a925:46397 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/932d5e8a06ee4ac5be223864979964f1 is 82, key is hbase:meta,,1/info:regioninfo/1733160420873/Put/seqid=0 2024-12-02T17:27:29,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741848_1033 (size=5672) 2024-12-02T17:27:29,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741848_1033 (size=5672) 2024-12-02T17:27:29,905 INFO [M:0;3b3375d1a925:46397 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=56 (bloomFilter=true), to=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/932d5e8a06ee4ac5be223864979964f1 2024-12-02T17:27:29,923 DEBUG [M:0;3b3375d1a925:46397 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/8a49c033baef408395eb40e202751c73 is 778, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733160421351/Put/seqid=0 2024-12-02T17:27:29,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741849_1034 (size=6118) 2024-12-02T17:27:29,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741849_1034 (size=6118) 2024-12-02T17:27:29,930 INFO [M:0;3b3375d1a925:46397 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=22.57 KB at sequenceid=56 (bloomFilter=true), to=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/8a49c033baef408395eb40e202751c73 2024-12-02T17:27:29,948 DEBUG [M:0;3b3375d1a925:46397 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/5101e997d5674757a67a114329898170 is 69, key is 3b3375d1a925,46459,1733160420055/rs:state/1733160420295/Put/seqid=0 2024-12-02T17:27:29,952 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741850_1035 (size=5156) 2024-12-02T17:27:29,953 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741850_1035 (size=5156) 2024-12-02T17:27:29,953 INFO [M:0;3b3375d1a925:46397 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=56 (bloomFilter=true), to=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/5101e997d5674757a67a114329898170 2024-12-02T17:27:29,962 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:27:29,962 INFO [RS:0;3b3375d1a925:46459 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:27:29,963 INFO [RS:0;3b3375d1a925:46459 {}] regionserver.HRegionServer(1031): Exiting; stopping=3b3375d1a925,46459,1733160420055; zookeeper connection closed. 2024-12-02T17:27:29,963 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46459-0x1009c074d360001, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:27:29,963 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@473c59de {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@473c59de 2024-12-02T17:27:29,963 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T17:27:29,971 DEBUG [M:0;3b3375d1a925:46397 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/6338c11035dd4448ad625ed94753d74b is 52, key is load_balancer_on/state:d/1733160420981/Put/seqid=0 2024-12-02T17:27:29,975 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741851_1036 (size=5056) 2024-12-02T17:27:29,975 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741851_1036 (size=5056) 2024-12-02T17:27:29,976 INFO [M:0;3b3375d1a925:46397 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=56 (bloomFilter=true), to=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/6338c11035dd4448ad625ed94753d74b 2024-12-02T17:27:29,981 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/932d5e8a06ee4ac5be223864979964f1 as hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/932d5e8a06ee4ac5be223864979964f1 2024-12-02T17:27:29,986 INFO [M:0;3b3375d1a925:46397 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/932d5e8a06ee4ac5be223864979964f1, entries=8, sequenceid=56, filesize=5.5 K 2024-12-02T17:27:29,987 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/8a49c033baef408395eb40e202751c73 as hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/8a49c033baef408395eb40e202751c73 2024-12-02T17:27:29,988 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:27:29,994 INFO [M:0;3b3375d1a925:46397 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/8a49c033baef408395eb40e202751c73, entries=6, sequenceid=56, filesize=6.0 K 2024-12-02T17:27:29,995 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/5101e997d5674757a67a114329898170 as hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/5101e997d5674757a67a114329898170 2024-12-02T17:27:29,999 INFO [M:0;3b3375d1a925:46397 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/5101e997d5674757a67a114329898170, entries=1, sequenceid=56, filesize=5.0 K 2024-12-02T17:27:30,000 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/6338c11035dd4448ad625ed94753d74b as hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/6338c11035dd4448ad625ed94753d74b 2024-12-02T17:27:30,006 INFO [M:0;3b3375d1a925:46397 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/6338c11035dd4448ad625ed94753d74b, entries=1, sequenceid=56, filesize=4.9 K 2024-12-02T17:27:30,007 INFO [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(3140): Finished flush of dataSize ~23.17 KB/23726, heapSize ~29.10 KB/29800, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 131ms, sequenceid=56, compaction requested=false 2024-12-02T17:27:30,008 INFO [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:30,008 DEBUG [M:0;3b3375d1a925:46397 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160449876Disabling compacts and flushes for region at 1733160449876Disabling writes for close at 1733160449876Obtaining lock to block concurrent updates at 1733160449876Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733160449876Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=23726, getHeapSize=29800, getOffHeapSize=0, getCellsCount=67 at 1733160449877 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733160449885 (+8 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733160449885Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733160449899 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733160449899Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733160449910 (+11 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733160449923 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733160449923Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733160449935 (+12 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733160449948 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733160449948Flushing 1595e783b53d99cd5eef43b6debb2682/state: creating writer at 1733160449958 (+10 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: appending metadata at 1733160449971 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: closing flushed file at 1733160449971Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5f81bb10: reopening flushed file at 1733160449981 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@44099af4: reopening flushed file at 1733160449986 (+5 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@46eb2df7: reopening flushed file at 1733160449994 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5f009e91: reopening flushed file at 1733160450000 (+6 ms)Finished flush of dataSize ~23.17 KB/23726, heapSize ~29.10 KB/29800, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 131ms, sequenceid=56, compaction requested=false at 1733160450007 (+7 ms)Writing region close event to WAL at 1733160450008 (+1 ms)Closed at 1733160450008 2024-12-02T17:27:30,009 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:30,009 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:30,009 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:30,009 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:30,009 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:27:30,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42681 is added to blk_1073741847_1031 (size=757) 2024-12-02T17:27:30,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37039 is added to blk_1073741847_1031 (size=757) 2024-12-02T17:27:30,998 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:30,998 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,009 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,009 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,009 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,009 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,010 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,010 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,012 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,013 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,013 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,015 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,019 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,020 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,522 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:27:31,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,536 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,537 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,537 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,537 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,537 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,538 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,541 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,541 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,542 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:31,543 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:32,860 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741830_1015: GenerationStamp not matched, existing replica is blk_1073741830_1006 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-02T17:27:33,886 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 after 4001ms 2024-12-02T17:27:33,886 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/WALs/3b3375d1a925,46397,1733160420014/3b3375d1a925%2C46397%2C1733160420014.1733160420146 to hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/oldWALs/3b3375d1a925%2C46397%2C1733160420014.1733160420146 2024-12-02T17:27:33,890 INFO [WAL-Archive-0 {}] region.MasterRegionUtils(50): Moved hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/MasterData/oldWALs/3b3375d1a925%2C46397%2C1733160420014.1733160420146 to hdfs://localhost:33813/user/jenkins/test-data/145424c9-7276-7f69-d13c-3a0937086009/oldWALs/3b3375d1a925%2C46397%2C1733160420014.1733160420146$masterlocalwal$ 2024-12-02T17:27:33,890 INFO [M:0;3b3375d1a925:46397 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T17:27:33,890 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:27:33,890 INFO [M:0;3b3375d1a925:46397 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:46397 2024-12-02T17:27:33,890 INFO [M:0;3b3375d1a925:46397 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:27:33,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:27:33,993 INFO [M:0;3b3375d1a925:46397 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:27:33,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46397-0x1009c074d360000, quorum=127.0.0.1:61144, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:27:33,995 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1ed6e4b8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:33,996 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@929fb8a{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:27:33,996 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:27:33,996 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3172a7c6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:27:33,996 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@652ca842{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,STOPPED} 2024-12-02T17:27:33,997 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:27:33,997 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:27:33,997 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1459169980-172.17.0.3-1733160419308 (Datanode Uuid 3ab70b2f-12bf-43be-9223-19d2e4e262b2) service to localhost/127.0.0.1:33813 2024-12-02T17:27:33,997 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:27:33,998 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data3/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:33,998 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data4/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:33,998 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:27:34,000 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5036ecf5{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:34,000 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@2647e756{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:27:34,001 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:27:34,001 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@290ac13e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:27:34,001 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@67d70e61{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,STOPPED} 2024-12-02T17:27:34,002 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:27:34,002 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:27:34,002 WARN [BP-1459169980-172.17.0.3-1733160419308 heartbeating to localhost/127.0.0.1:33813 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1459169980-172.17.0.3-1733160419308 (Datanode Uuid c667f086-f4c3-4afc-ae83-87fcf18d5c75) service to localhost/127.0.0.1:33813 2024-12-02T17:27:34,002 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:27:34,002 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data1/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:34,003 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/cluster_cd456194-267d-f63d-f03f-112ff1876fb4/data/data2/current/BP-1459169980-172.17.0.3-1733160419308 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:27:34,003 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:27:34,014 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7e54696a{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:27:34,015 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@228e200c{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:27:34,015 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:27:34,015 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@65e5bfc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:27:34,015 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@affd6eb{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir/,STOPPED} 2024-12-02T17:27:34,021 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T17:27:34,038 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T17:27:34,049 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnPipelineRestart Thread=177 (was 152) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-14 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-32-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33813 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-30-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-10-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-11-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-13 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-30-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-31-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-33-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-31-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-10-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-12 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:33813 from jenkins.hfs.4 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: LeaseRenewer:jenkins@localhost:33813 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-33-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-30-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33813 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-15 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:33813 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33813 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-11-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-32-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-10-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-32-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:33813 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-33-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-11-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-31-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.4@localhost:33813 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=457 (was 448) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=178 (was 220), ProcessCount=11 (was 11), AvailableMemoryMB=2146 (was 2300) 2024-12-02T17:27:34,057 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testCompactionRecordDoesntBlockRolling Thread=177, OpenFileDescriptor=457, MaxFileDescriptor=1048576, SystemLoadAverage=178, ProcessCount=11, AvailableMemoryMB=2147 2024-12-02T17:27:34,057 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T17:27:34,057 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.log.dir so I do NOT create it in target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd 2024-12-02T17:27:34,057 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c9d79e1e-e75c-59aa-07e0-3cdf7f5b308e/hadoop.tmp.dir so I do NOT create it in target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd 2024-12-02T17:27:34,057 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c, deleteOnExit=true 2024-12-02T17:27:34,057 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-02T17:27:34,057 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/test.cache.data in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.log.dir in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T17:27:34,058 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T17:27:34,058 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T17:27:34,059 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:27:34,059 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:27:34,059 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T17:27:34,059 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/nfs.dump.dir in system properties and HBase conf 2024-12-02T17:27:34,059 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/java.io.tmpdir in system properties and HBase conf 2024-12-02T17:27:34,059 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:27:34,059 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T17:27:34,059 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T17:27:34,071 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:27:34,135 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:27:34,139 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:27:34,140 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:27:34,140 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:27:34,141 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:27:34,142 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:27:34,144 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@15cd018{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:27:34,145 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4a31a089{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:27:34,269 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@47df503f{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/java.io.tmpdir/jetty-localhost-37671-hadoop-hdfs-3_4_1-tests_jar-_-any-4684182517368953012/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:27:34,269 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5f3519da{HTTP/1.1, (http/1.1)}{localhost:37671} 2024-12-02T17:27:34,269 INFO [Time-limited test {}] server.Server(415): Started @181942ms 2024-12-02T17:27:34,281 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:27:34,334 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:27:34,336 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:27:34,337 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:27:34,337 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:27:34,337 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:27:34,337 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@25c02940{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:27:34,338 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5718d675{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:27:34,451 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1af676f5{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/java.io.tmpdir/jetty-localhost-38891-hadoop-hdfs-3_4_1-tests_jar-_-any-17164070925516168835/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:34,452 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@9cd716d{HTTP/1.1, (http/1.1)}{localhost:38891} 2024-12-02T17:27:34,452 INFO [Time-limited test {}] server.Server(415): Started @182124ms 2024-12-02T17:27:34,453 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:27:34,482 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:27:34,484 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:27:34,485 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:27:34,485 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:27:34,485 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:27:34,485 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6df2cf02{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:27:34,485 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4dfe8781{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:27:34,599 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@191911fe{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/java.io.tmpdir/jetty-localhost-41223-hadoop-hdfs-3_4_1-tests_jar-_-any-5113832912103375260/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:27:34,599 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@30d9f702{HTTP/1.1, (http/1.1)}{localhost:41223} 2024-12-02T17:27:34,599 INFO [Time-limited test {}] server.Server(415): Started @182271ms 2024-12-02T17:27:34,600 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:27:34,729 WARN [Thread-1636 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/data/data2/current/BP-1826006310-172.17.0.3-1733160454087/current, will proceed with Du for space computation calculation, 2024-12-02T17:27:34,729 WARN [Thread-1635 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/data/data1/current/BP-1826006310-172.17.0.3-1733160454087/current, will proceed with Du for space computation calculation, 2024-12-02T17:27:34,752 WARN [Thread-1601 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:27:34,754 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdba140edac039358 with lease ID 0x702263c92dbefdd4: Processing first storage report for DS-8dc1a0a2-db36-4e86-a408-f0ee7136a916 from datanode DatanodeRegistration(127.0.0.1:45461, datanodeUuid=e4714d92-0927-492d-8fb0-ba7f5f536e47, infoPort=46083, infoSecurePort=0, ipcPort=36289, storageInfo=lv=-57;cid=testClusterID;nsid=953323972;c=1733160454087) 2024-12-02T17:27:34,754 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdba140edac039358 with lease ID 0x702263c92dbefdd4: from storage DS-8dc1a0a2-db36-4e86-a408-f0ee7136a916 node DatanodeRegistration(127.0.0.1:45461, datanodeUuid=e4714d92-0927-492d-8fb0-ba7f5f536e47, infoPort=46083, infoSecurePort=0, ipcPort=36289, storageInfo=lv=-57;cid=testClusterID;nsid=953323972;c=1733160454087), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:34,754 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdba140edac039358 with lease ID 0x702263c92dbefdd4: Processing first storage report for DS-b4541b64-43f5-416b-871b-5efd6df0ebbe from datanode DatanodeRegistration(127.0.0.1:45461, datanodeUuid=e4714d92-0927-492d-8fb0-ba7f5f536e47, infoPort=46083, infoSecurePort=0, ipcPort=36289, storageInfo=lv=-57;cid=testClusterID;nsid=953323972;c=1733160454087) 2024-12-02T17:27:34,754 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdba140edac039358 with lease ID 0x702263c92dbefdd4: from storage DS-b4541b64-43f5-416b-871b-5efd6df0ebbe node DatanodeRegistration(127.0.0.1:45461, datanodeUuid=e4714d92-0927-492d-8fb0-ba7f5f536e47, infoPort=46083, infoSecurePort=0, ipcPort=36289, storageInfo=lv=-57;cid=testClusterID;nsid=953323972;c=1733160454087), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:34,791 WARN [Thread-1648 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/data/data3/current/BP-1826006310-172.17.0.3-1733160454087/current, will proceed with Du for space computation calculation, 2024-12-02T17:27:34,791 WARN [Thread-1649 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/data/data4/current/BP-1826006310-172.17.0.3-1733160454087/current, will proceed with Du for space computation calculation, 2024-12-02T17:27:34,807 WARN [Thread-1625 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:27:34,809 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc1a52f4ec39c5aa6 with lease ID 0x702263c92dbefdd5: Processing first storage report for DS-e7110b49-f6f8-44f3-be94-35a71c8d745b from datanode DatanodeRegistration(127.0.0.1:39997, datanodeUuid=783409ae-4421-47a1-9609-9b02818e9a25, infoPort=35973, infoSecurePort=0, ipcPort=41691, storageInfo=lv=-57;cid=testClusterID;nsid=953323972;c=1733160454087) 2024-12-02T17:27:34,809 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc1a52f4ec39c5aa6 with lease ID 0x702263c92dbefdd5: from storage DS-e7110b49-f6f8-44f3-be94-35a71c8d745b node DatanodeRegistration(127.0.0.1:39997, datanodeUuid=783409ae-4421-47a1-9609-9b02818e9a25, infoPort=35973, infoSecurePort=0, ipcPort=41691, storageInfo=lv=-57;cid=testClusterID;nsid=953323972;c=1733160454087), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:34,809 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc1a52f4ec39c5aa6 with lease ID 0x702263c92dbefdd5: Processing first storage report for DS-94337727-3bce-4353-99c1-3b4790b8dabc from datanode DatanodeRegistration(127.0.0.1:39997, datanodeUuid=783409ae-4421-47a1-9609-9b02818e9a25, infoPort=35973, infoSecurePort=0, ipcPort=41691, storageInfo=lv=-57;cid=testClusterID;nsid=953323972;c=1733160454087) 2024-12-02T17:27:34,809 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc1a52f4ec39c5aa6 with lease ID 0x702263c92dbefdd5: from storage DS-94337727-3bce-4353-99c1-3b4790b8dabc node DatanodeRegistration(127.0.0.1:39997, datanodeUuid=783409ae-4421-47a1-9609-9b02818e9a25, infoPort=35973, infoSecurePort=0, ipcPort=41691, storageInfo=lv=-57;cid=testClusterID;nsid=953323972;c=1733160454087), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:27:34,820 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd 2024-12-02T17:27:34,823 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/zookeeper_0, clientPort=63626, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T17:27:34,823 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=63626 2024-12-02T17:27:34,824 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:34,825 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:34,832 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:27:34,833 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:27:34,833 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32 with version=8 2024-12-02T17:27:34,833 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase-staging 2024-12-02T17:27:34,835 INFO [Time-limited test {}] client.ConnectionUtils(128): master/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:27:34,835 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:34,835 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:34,835 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:27:34,835 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:34,835 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:27:34,835 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T17:27:34,836 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:27:34,836 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:46201 2024-12-02T17:27:34,837 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:46201 connecting to ZooKeeper ensemble=127.0.0.1:63626 2024-12-02T17:27:34,843 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:462010x0, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:27:34,843 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:46201-0x1009c07d5450000 connected 2024-12-02T17:27:34,856 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:34,857 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:34,859 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:27:34,859 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32, hbase.cluster.distributed=false 2024-12-02T17:27:34,860 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:27:34,861 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46201 2024-12-02T17:27:34,861 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46201 2024-12-02T17:27:34,861 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46201 2024-12-02T17:27:34,861 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46201 2024-12-02T17:27:34,862 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46201 2024-12-02T17:27:34,876 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:27:34,876 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:34,876 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:34,876 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:27:34,876 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:27:34,876 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:27:34,876 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T17:27:34,876 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:27:34,877 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:33741 2024-12-02T17:27:34,878 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:33741 connecting to ZooKeeper ensemble=127.0.0.1:63626 2024-12-02T17:27:34,878 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:34,880 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:34,885 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:337410x0, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:27:34,886 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:337410x0, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:27:34,886 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:33741-0x1009c07d5450001 connected 2024-12-02T17:27:34,886 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T17:27:34,886 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T17:27:34,887 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T17:27:34,888 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:27:34,888 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=33741 2024-12-02T17:27:34,888 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=33741 2024-12-02T17:27:34,889 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=33741 2024-12-02T17:27:34,889 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=33741 2024-12-02T17:27:34,889 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=33741 2024-12-02T17:27:34,900 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;3b3375d1a925:46201 2024-12-02T17:27:34,900 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/3b3375d1a925,46201,1733160454835 2024-12-02T17:27:34,902 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:27:34,902 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:27:34,902 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/3b3375d1a925,46201,1733160454835 2024-12-02T17:27:34,904 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T17:27:34,904 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:34,904 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:34,904 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T17:27:34,905 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/3b3375d1a925,46201,1733160454835 from backup master directory 2024-12-02T17:27:34,906 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/3b3375d1a925,46201,1733160454835 2024-12-02T17:27:34,906 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:27:34,906 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:27:34,906 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:27:34,906 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=3b3375d1a925,46201,1733160454835 2024-12-02T17:27:34,910 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/hbase.id] with ID: 68848a20-adda-42d9-8645-152adafce4ef 2024-12-02T17:27:34,910 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/.tmp/hbase.id 2024-12-02T17:27:34,914 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:27:34,915 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:27:34,915 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/.tmp/hbase.id]:[hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/hbase.id] 2024-12-02T17:27:34,926 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:34,926 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T17:27:34,927 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 1ms. 2024-12-02T17:27:34,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:34,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:34,935 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:27:34,935 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:27:34,936 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:27:34,937 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T17:27:34,940 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:27:34,949 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:27:34,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:27:34,951 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store 2024-12-02T17:27:34,962 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:27:34,962 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:27:34,964 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:34,964 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:27:34,964 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:34,964 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:34,964 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:27:34,964 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:34,964 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:27:34,964 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160454964Disabling compacts and flushes for region at 1733160454964Disabling writes for close at 1733160454964Writing region close event to WAL at 1733160454964Closed at 1733160454964 2024-12-02T17:27:34,965 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/.initializing 2024-12-02T17:27:34,965 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/WALs/3b3375d1a925,46201,1733160454835 2024-12-02T17:27:34,968 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C46201%2C1733160454835, suffix=, logDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/WALs/3b3375d1a925,46201,1733160454835, archiveDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/oldWALs, maxLogs=10 2024-12-02T17:27:34,968 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C46201%2C1733160454835.1733160454968 2024-12-02T17:27:34,977 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/WALs/3b3375d1a925,46201,1733160454835/3b3375d1a925%2C46201%2C1733160454835.1733160454968 2024-12-02T17:27:34,982 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46083:46083),(127.0.0.1/127.0.0.1:35973:35973)] 2024-12-02T17:27:34,986 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:27:34,986 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:34,986 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,986 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,989 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,990 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T17:27:34,990 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:34,991 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:34,991 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,992 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T17:27:34,992 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:34,992 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:27:34,993 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,994 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T17:27:34,994 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:34,994 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:27:34,994 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,995 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T17:27:34,995 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:34,996 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:27:34,996 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,997 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,997 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,999 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,999 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:34,999 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T17:27:35,001 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:27:35,003 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:27:35,003 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=879682, jitterRate=0.11857366561889648}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T17:27:35,004 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733160454986Initializing all the Stores at 1733160454987 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160454987Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160454988 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160454988Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160454988Cleaning up temporary data from old regions at 1733160454999 (+11 ms)Region opened successfully at 1733160455004 (+5 ms) 2024-12-02T17:27:35,004 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T17:27:35,007 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@59f44402, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:27:35,008 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T17:27:35,008 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T17:27:35,008 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T17:27:35,008 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T17:27:35,009 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T17:27:35,009 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-02T17:27:35,009 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T17:27:35,011 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T17:27:35,012 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T17:27:35,013 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T17:27:35,014 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T17:27:35,014 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T17:27:35,015 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T17:27:35,016 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T17:27:35,017 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T17:27:35,018 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T17:27:35,019 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T17:27:35,020 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T17:27:35,022 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T17:27:35,023 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T17:27:35,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:27:35,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:35,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:27:35,025 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:35,025 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=3b3375d1a925,46201,1733160454835, sessionid=0x1009c07d5450000, setting cluster-up flag (Was=false) 2024-12-02T17:27:35,028 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:35,028 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:35,033 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T17:27:35,034 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,46201,1733160454835 2024-12-02T17:27:35,037 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:35,037 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:35,043 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T17:27:35,044 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,46201,1733160454835 2024-12-02T17:27:35,045 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T17:27:35,046 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T17:27:35,047 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T17:27:35,047 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T17:27:35,047 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 3b3375d1a925,46201,1733160454835 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T17:27:35,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:27:35,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:27:35,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:27:35,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:27:35,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/3b3375d1a925:0, corePoolSize=10, maxPoolSize=10 2024-12-02T17:27:35,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:27:35,048 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,052 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:27:35,052 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733160485052 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T17:27:35,053 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,053 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T17:27:35,053 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T17:27:35,054 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T17:27:35,054 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T17:27:35,054 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T17:27:35,054 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160455054,5,FailOnTimeoutGroup] 2024-12-02T17:27:35,054 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160455054,5,FailOnTimeoutGroup] 2024-12-02T17:27:35,054 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,054 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T17:27:35,054 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,054 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,063 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:27:35,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:27:35,064 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T17:27:35,065 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32 2024-12-02T17:27:35,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:27:35,074 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:27:35,074 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:35,076 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:27:35,078 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:27:35,078 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,078 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:35,078 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:27:35,079 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:27:35,079 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,079 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:35,080 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:27:35,081 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:27:35,081 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,081 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:35,081 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:27:35,082 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:27:35,082 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,083 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:35,083 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:27:35,083 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740 2024-12-02T17:27:35,084 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740 2024-12-02T17:27:35,085 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:27:35,085 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:27:35,086 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:27:35,087 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:27:35,089 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:27:35,089 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=779109, jitterRate=-0.00931212306022644}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:27:35,090 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733160455074Initializing all the Stores at 1733160455075 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160455075Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160455076 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160455076Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160455076Cleaning up temporary data from old regions at 1733160455085 (+9 ms)Region opened successfully at 1733160455090 (+5 ms) 2024-12-02T17:27:35,090 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:27:35,090 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:27:35,090 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:27:35,090 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:27:35,090 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:27:35,091 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:27:35,091 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160455090Disabling compacts and flushes for region at 1733160455090Disabling writes for close at 1733160455090Writing region close event to WAL at 1733160455091 (+1 ms)Closed at 1733160455091 2024-12-02T17:27:35,091 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(746): ClusterId : 68848a20-adda-42d9-8645-152adafce4ef 2024-12-02T17:27:35,091 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T17:27:35,092 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:27:35,092 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T17:27:35,092 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T17:27:35,093 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T17:27:35,093 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T17:27:35,094 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:27:35,095 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T17:27:35,097 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T17:27:35,097 DEBUG [RS:0;3b3375d1a925:33741 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1098e92, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:27:35,108 DEBUG [RS:0;3b3375d1a925:33741 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;3b3375d1a925:33741 2024-12-02T17:27:35,109 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T17:27:35,109 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T17:27:35,109 DEBUG [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T17:27:35,109 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(2659): reportForDuty to master=3b3375d1a925,46201,1733160454835 with port=33741, startcode=1733160454876 2024-12-02T17:27:35,109 DEBUG [RS:0;3b3375d1a925:33741 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T17:27:35,111 INFO [HMaster-EventLoopGroup-12-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43511, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T17:27:35,112 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46201 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,112 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46201 {}] master.ServerManager(517): Registering regionserver=3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,114 DEBUG [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32 2024-12-02T17:27:35,114 DEBUG [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42485 2024-12-02T17:27:35,114 DEBUG [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T17:27:35,115 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:27:35,116 DEBUG [RS:0;3b3375d1a925:33741 {}] zookeeper.ZKUtil(111): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,116 WARN [RS:0;3b3375d1a925:33741 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:27:35,116 INFO [RS:0;3b3375d1a925:33741 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:27:35,116 DEBUG [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,116 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [3b3375d1a925,33741,1733160454876] 2024-12-02T17:27:35,122 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T17:27:35,124 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T17:27:35,124 INFO [RS:0;3b3375d1a925:33741 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:27:35,124 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,124 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T17:27:35,125 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T17:27:35,125 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,125 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,125 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,125 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,125 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,125 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:27:35,126 DEBUG [RS:0;3b3375d1a925:33741 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:27:35,127 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,127 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,127 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,127 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,127 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,127 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,33741,1733160454876-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:27:35,142 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T17:27:35,142 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,33741,1733160454876-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,142 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,142 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.Replication(171): 3b3375d1a925,33741,1733160454876 started 2024-12-02T17:27:35,156 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,156 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1482): Serving as 3b3375d1a925,33741,1733160454876, RpcServer on 3b3375d1a925/172.17.0.3:33741, sessionid=0x1009c07d5450001 2024-12-02T17:27:35,156 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T17:27:35,156 DEBUG [RS:0;3b3375d1a925:33741 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,156 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,33741,1733160454876' 2024-12-02T17:27:35,156 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T17:27:35,156 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T17:27:35,157 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T17:27:35,157 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T17:27:35,157 DEBUG [RS:0;3b3375d1a925:33741 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,157 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,33741,1733160454876' 2024-12-02T17:27:35,157 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T17:27:35,157 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T17:27:35,158 DEBUG [RS:0;3b3375d1a925:33741 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T17:27:35,158 INFO [RS:0;3b3375d1a925:33741 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T17:27:35,158 INFO [RS:0;3b3375d1a925:33741 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T17:27:35,245 WARN [3b3375d1a925:46201 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-02T17:27:35,259 INFO [RS:0;3b3375d1a925:33741 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C33741%2C1733160454876, suffix=, logDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876, archiveDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/oldWALs, maxLogs=32 2024-12-02T17:27:35,260 INFO [RS:0;3b3375d1a925:33741 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C33741%2C1733160454876.1733160455260 2024-12-02T17:27:35,265 INFO [RS:0;3b3375d1a925:33741 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160455260 2024-12-02T17:27:35,266 DEBUG [RS:0;3b3375d1a925:33741 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35973:35973),(127.0.0.1/127.0.0.1:46083:46083)] 2024-12-02T17:27:35,495 DEBUG [3b3375d1a925:46201 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T17:27:35,496 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,497 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,33741,1733160454876, state=OPENING 2024-12-02T17:27:35,499 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T17:27:35,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:35,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:27:35,501 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:27:35,501 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:27:35,501 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:27:35,501 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,33741,1733160454876}] 2024-12-02T17:27:35,654 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T17:27:35,656 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-13-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37775, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T17:27:35,659 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T17:27:35,659 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:27:35,661 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C33741%2C1733160454876.meta, suffix=.meta, logDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876, archiveDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/oldWALs, maxLogs=32 2024-12-02T17:27:35,661 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C33741%2C1733160454876.meta.1733160455661.meta 2024-12-02T17:27:35,666 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.meta.1733160455661.meta 2024-12-02T17:27:35,667 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46083:46083),(127.0.0.1/127.0.0.1:35973:35973)] 2024-12-02T17:27:35,667 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:27:35,668 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T17:27:35,668 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T17:27:35,668 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T17:27:35,668 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T17:27:35,668 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:35,668 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T17:27:35,668 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T17:27:35,669 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:27:35,670 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:27:35,670 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,671 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:35,671 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:27:35,671 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:27:35,672 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,672 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:35,672 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:27:35,673 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:27:35,673 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,673 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:35,673 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:27:35,674 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:27:35,674 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,674 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:27:35,674 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:27:35,675 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740 2024-12-02T17:27:35,675 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740 2024-12-02T17:27:35,676 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:27:35,677 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:27:35,677 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:27:35,678 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:27:35,679 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=867424, jitterRate=0.10298667848110199}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:27:35,679 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T17:27:35,679 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733160455668Writing region info on filesystem at 1733160455668Initializing all the Stores at 1733160455669 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160455669Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160455669Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160455669Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160455669Cleaning up temporary data from old regions at 1733160455677 (+8 ms)Running coprocessor post-open hooks at 1733160455679 (+2 ms)Region opened successfully at 1733160455679 2024-12-02T17:27:35,680 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733160455654 2024-12-02T17:27:35,682 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T17:27:35,682 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T17:27:35,683 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,684 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,33741,1733160454876, state=OPEN 2024-12-02T17:27:35,688 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:27:35,688 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:27:35,688 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,688 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:27:35,688 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:27:35,691 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T17:27:35,691 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,33741,1733160454876 in 187 msec 2024-12-02T17:27:35,693 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T17:27:35,693 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 599 msec 2024-12-02T17:27:35,694 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:27:35,694 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T17:27:35,695 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:27:35,695 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,33741,1733160454876, seqNum=-1] 2024-12-02T17:27:35,695 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:27:35,697 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-13-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:60639, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:27:35,701 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 655 msec 2024-12-02T17:27:35,702 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733160455702, completionTime=-1 2024-12-02T17:27:35,702 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T17:27:35,702 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T17:27:35,703 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=1 2024-12-02T17:27:35,704 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733160515704 2024-12-02T17:27:35,704 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733160575704 2024-12-02T17:27:35,704 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 1 msec 2024-12-02T17:27:35,704 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46201,1733160454835-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,704 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46201,1733160454835-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,704 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46201,1733160454835-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,704 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-3b3375d1a925:46201, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,704 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,706 DEBUG [master/3b3375d1a925:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T17:27:35,707 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,710 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.804sec 2024-12-02T17:27:35,710 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T17:27:35,710 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T17:27:35,710 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T17:27:35,710 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T17:27:35,710 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T17:27:35,710 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46201,1733160454835-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:27:35,710 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46201,1733160454835-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T17:27:35,712 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T17:27:35,712 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T17:27:35,712 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,46201,1733160454835-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:27:35,791 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@65c292b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:27:35,791 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 3b3375d1a925,46201,-1 for getting cluster id 2024-12-02T17:27:35,791 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T17:27:35,793 DEBUG [HMaster-EventLoopGroup-12-3 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '68848a20-adda-42d9-8645-152adafce4ef' 2024-12-02T17:27:35,793 DEBUG [RPCClient-NioEventLoopGroup-4-16 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T17:27:35,793 DEBUG [RPCClient-NioEventLoopGroup-4-16 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "68848a20-adda-42d9-8645-152adafce4ef" 2024-12-02T17:27:35,793 DEBUG [RPCClient-NioEventLoopGroup-4-16 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6de3e0cb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:27:35,793 DEBUG [RPCClient-NioEventLoopGroup-4-16 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [3b3375d1a925,46201,-1] 2024-12-02T17:27:35,793 DEBUG [RPCClient-NioEventLoopGroup-4-16 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T17:27:35,794 DEBUG [RPCClient-NioEventLoopGroup-4-16 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:27:35,795 INFO [HMaster-EventLoopGroup-12-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45670, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T17:27:35,795 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@34fe4f6c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:27:35,796 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:27:35,797 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,33741,1733160454876, seqNum=-1] 2024-12-02T17:27:35,797 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:27:35,798 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-13-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51720, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:27:35,799 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=3b3375d1a925,46201,1733160454835 2024-12-02T17:27:35,799 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:27:35,802 INFO [Time-limited test {}] master.MasterRpcServices(567): Client=null/null set balanceSwitch=false 2024-12-02T17:27:35,802 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-02T17:27:35,803 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] client.AsyncConnectionImpl(321): The fetched master address is 3b3375d1a925,46201,1733160454835 2024-12-02T17:27:35,803 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@307a19aa 2024-12-02T17:27:35,803 DEBUG [RPCClient-NioEventLoopGroup-4-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T17:27:35,804 INFO [HMaster-EventLoopGroup-12-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45682, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T17:27:35,804 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T17:27:35,804 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T17:27:35,804 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestLogRolling-testCompactionRecordDoesntBlockRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:27:35,805 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:27:35,806 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T17:27:35,807 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:35,807 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testCompactionRecordDoesntBlockRolling" procId is: 4 2024-12-02T17:27:35,807 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:27:35,808 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T17:27:35,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741835_1011 (size=405) 2024-12-02T17:27:35,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741835_1011 (size=405) 2024-12-02T17:27:35,815 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 8c1b9a0082d5e51306447fb7be8c0154, NAME => 'TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testCompactionRecordDoesntBlockRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32 2024-12-02T17:27:35,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741836_1012 (size=88) 2024-12-02T17:27:35,821 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741836_1012 (size=88) 2024-12-02T17:27:35,821 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(898): Instantiated TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:35,821 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1722): Closing 8c1b9a0082d5e51306447fb7be8c0154, disabling compactions & flushes 2024-12-02T17:27:35,821 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1755): Closing region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:35,822 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:35,822 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. after waiting 0 ms 2024-12-02T17:27:35,822 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:35,822 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1973): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:35,822 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1676): Region close journal for 8c1b9a0082d5e51306447fb7be8c0154: Waiting for close lock at 1733160455821Disabling compacts and flushes for region at 1733160455821Disabling writes for close at 1733160455822 (+1 ms)Writing region close event to WAL at 1733160455822Closed at 1733160455822 2024-12-02T17:27:35,823 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T17:27:35,823 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.","families":{"info":[{"qualifier":"regioninfo","vlen":87,"tag":[],"timestamp":"1733160455823"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733160455823"}]},"ts":"1733160455823"} 2024-12-02T17:27:35,825 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-02T17:27:35,826 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T17:27:35,826 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160455826"}]},"ts":"1733160455826"} 2024-12-02T17:27:35,828 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testCompactionRecordDoesntBlockRolling, state=ENABLING in hbase:meta 2024-12-02T17:27:35,829 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=8c1b9a0082d5e51306447fb7be8c0154, ASSIGN}] 2024-12-02T17:27:35,830 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=8c1b9a0082d5e51306447fb7be8c0154, ASSIGN 2024-12-02T17:27:35,831 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=8c1b9a0082d5e51306447fb7be8c0154, ASSIGN; state=OFFLINE, location=3b3375d1a925,33741,1733160454876; forceNewPlan=false, retain=false 2024-12-02T17:27:35,981 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=8c1b9a0082d5e51306447fb7be8c0154, regionState=OPENING, regionLocation=3b3375d1a925,33741,1733160454876 2024-12-02T17:27:35,984 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-13-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=8c1b9a0082d5e51306447fb7be8c0154, ASSIGN because future has completed 2024-12-02T17:27:35,984 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 8c1b9a0082d5e51306447fb7be8c0154, server=3b3375d1a925,33741,1733160454876}] 2024-12-02T17:27:36,141 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:36,141 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 8c1b9a0082d5e51306447fb7be8c0154, NAME => 'TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:27:36,141 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testCompactionRecordDoesntBlockRolling 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,141 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:27:36,141 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,141 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,143 INFO [StoreOpener-8c1b9a0082d5e51306447fb7be8c0154-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,144 INFO [StoreOpener-8c1b9a0082d5e51306447fb7be8c0154-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 8c1b9a0082d5e51306447fb7be8c0154 columnFamilyName info 2024-12-02T17:27:36,144 DEBUG [StoreOpener-8c1b9a0082d5e51306447fb7be8c0154-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:27:36,144 INFO [StoreOpener-8c1b9a0082d5e51306447fb7be8c0154-1 {}] regionserver.HStore(327): Store=8c1b9a0082d5e51306447fb7be8c0154/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:27:36,145 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,145 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,145 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,146 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,146 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,147 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,149 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:27:36,150 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 8c1b9a0082d5e51306447fb7be8c0154; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=742385, jitterRate=-0.05600924789905548}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T17:27:36,150 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:27:36,150 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 8c1b9a0082d5e51306447fb7be8c0154: Running coprocessor pre-open hook at 1733160456142Writing region info on filesystem at 1733160456142Initializing all the Stores at 1733160456142Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160456142Cleaning up temporary data from old regions at 1733160456146 (+4 ms)Running coprocessor post-open hooks at 1733160456150 (+4 ms)Region opened successfully at 1733160456150 2024-12-02T17:27:36,151 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154., pid=6, masterSystemTime=1733160456137 2024-12-02T17:27:36,154 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:36,154 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:36,155 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=8c1b9a0082d5e51306447fb7be8c0154, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,33741,1733160454876 2024-12-02T17:27:36,157 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-13-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 8c1b9a0082d5e51306447fb7be8c0154, server=3b3375d1a925,33741,1733160454876 because future has completed 2024-12-02T17:27:36,161 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T17:27:36,161 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 8c1b9a0082d5e51306447fb7be8c0154, server=3b3375d1a925,33741,1733160454876 in 174 msec 2024-12-02T17:27:36,163 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T17:27:36,163 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=8c1b9a0082d5e51306447fb7be8c0154, ASSIGN in 332 msec 2024-12-02T17:27:36,164 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T17:27:36,164 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160456164"}]},"ts":"1733160456164"} 2024-12-02T17:27:36,166 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testCompactionRecordDoesntBlockRolling, state=ENABLED in hbase:meta 2024-12-02T17:27:36,167 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T17:27:36,169 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 363 msec 2024-12-02T17:27:37,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:27:37,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-02T17:27:37,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:27:37,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-02T17:27:37,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:27:37,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling Metrics about Tables on a single HBase RegionServer 2024-12-02T17:27:41,170 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:27:41,171 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,171 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,171 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,172 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,172 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,172 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,186 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,186 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,186 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,186 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,187 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,187 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,190 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,190 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,190 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,192 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:27:41,197 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T17:27:41,197 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testCompactionRecordDoesntBlockRolling' 2024-12-02T17:27:45,899 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:27:45,900 INFO [RPCClient-NioEventLoopGroup-4-3 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling completed 2024-12-02T17:27:45,900 DEBUG [Time-limited test {}] hbase.ClientMetaTableAccessor(255): Scanning META starting at row=TestLogRolling-testCompactionRecordDoesntBlockRolling,, stopping at row=TestLogRolling-testCompactionRecordDoesntBlockRolling ,, for max=2147483647 with caching=100 2024-12-02T17:27:45,903 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2234): Found 1 regions for table TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:27:45,903 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2240): firstRegionName=TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:45,906 DEBUG [RPCClient-NioEventLoopGroup-4-2 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestLogRolling-testCompactionRecordDoesntBlockRolling', row='row0001', locateType=CURRENT is [region=TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154., hostname=3b3375d1a925,33741,1733160454876, seqNum=2] 2024-12-02T17:27:45,913 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:27:45,918 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:27:45,919 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-02T17:27:45,920 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T17:27:45,920 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T17:27:45,922 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T17:27:46,082 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=33741 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-02T17:27:46,083 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:46,083 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 8c1b9a0082d5e51306447fb7be8c0154 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T17:27:46,099 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/5ea7b4cdbd874cafb117575d7d0eb932 is 1080, key is row0001/info:/1733160465907/Put/seqid=0 2024-12-02T17:27:46,104 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741837_1013 (size=6033) 2024-12-02T17:27:46,104 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741837_1013 (size=6033) 2024-12-02T17:27:46,105 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/5ea7b4cdbd874cafb117575d7d0eb932 2024-12-02T17:27:46,111 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/5ea7b4cdbd874cafb117575d7d0eb932 as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/5ea7b4cdbd874cafb117575d7d0eb932 2024-12-02T17:27:46,116 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/5ea7b4cdbd874cafb117575d7d0eb932, entries=1, sequenceid=5, filesize=5.9 K 2024-12-02T17:27:46,117 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 8c1b9a0082d5e51306447fb7be8c0154 in 34ms, sequenceid=5, compaction requested=false 2024-12-02T17:27:46,117 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 8c1b9a0082d5e51306447fb7be8c0154: 2024-12-02T17:27:46,117 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:46,118 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-02T17:27:46,120 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-02T17:27:46,124 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-02T17:27:46,124 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 200 msec 2024-12-02T17:27:46,127 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 210 msec 2024-12-02T17:27:55,940 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-02T17:27:55,940 INFO [RPCClient-NioEventLoopGroup-4-3 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling completed 2024-12-02T17:27:55,943 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:27:55,944 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] procedure2.ProcedureExecutor(1139): Stored pid=9, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=9, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:27:55,945 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=9 2024-12-02T17:27:55,945 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=9, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=9, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-02T17:27:55,947 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=9, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=9, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T17:27:55,947 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T17:27:56,100 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=33741 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=10 2024-12-02T17:27:56,100 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:56,100 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.HRegion(2902): Flushing 8c1b9a0082d5e51306447fb7be8c0154 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T17:27:56,105 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/19a7be03f1b74211ac10f6a6eb303545 is 1080, key is row0002/info:/1733160475941/Put/seqid=0 2024-12-02T17:27:56,109 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741838_1014 (size=6033) 2024-12-02T17:27:56,110 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741838_1014 (size=6033) 2024-12-02T17:27:56,110 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/19a7be03f1b74211ac10f6a6eb303545 2024-12-02T17:27:56,117 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/19a7be03f1b74211ac10f6a6eb303545 as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/19a7be03f1b74211ac10f6a6eb303545 2024-12-02T17:27:56,121 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/19a7be03f1b74211ac10f6a6eb303545, entries=1, sequenceid=9, filesize=5.9 K 2024-12-02T17:27:56,122 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.HRegion(3140): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 8c1b9a0082d5e51306447fb7be8c0154 in 22ms, sequenceid=9, compaction requested=false 2024-12-02T17:27:56,122 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.HRegion(2603): Flush status journal for 8c1b9a0082d5e51306447fb7be8c0154: 2024-12-02T17:27:56,123 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:27:56,123 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-1 {event_type=RS_FLUSH_REGIONS, pid=10}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=10 2024-12-02T17:27:56,123 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster(4169): Remote procedure done, pid=10 2024-12-02T17:27:56,127 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=10, resume processing ppid=9 2024-12-02T17:27:56,127 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1521): Finished pid=10, ppid=9, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 177 msec 2024-12-02T17:27:56,129 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=9, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=9, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 184 msec 2024-12-02T17:28:04,820 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:28:06,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=9 2024-12-02T17:28:06,020 INFO [RPCClient-NioEventLoopGroup-4-3 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling completed 2024-12-02T17:28:06,023 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C33741%2C1733160454876.1733160486023 2024-12-02T17:28:06,029 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:06,029 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:06,029 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:06,029 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:06,029 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:06,029 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160455260 with entries=8, filesize=5.41 KB; new WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160486023 2024-12-02T17:28:06,030 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46083:46083),(127.0.0.1/127.0.0.1:35973:35973)] 2024-12-02T17:28:06,030 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160455260 is not closed yet, will try archiving it next time 2024-12-02T17:28:06,031 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:28:06,031 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741833_1009 (size=5546) 2024-12-02T17:28:06,031 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741833_1009 (size=5546) 2024-12-02T17:28:06,032 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] procedure2.ProcedureExecutor(1139): Stored pid=11, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=11, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:28:06,033 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=11 2024-12-02T17:28:06,033 INFO [PEWorker-5 {}] procedure.FlushTableProcedure(91): pid=11, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=11, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-02T17:28:06,034 INFO [PEWorker-5 {}] procedure.FlushTableProcedure(91): pid=11, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=11, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T17:28:06,034 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=12, ppid=11, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T17:28:06,187 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=33741 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=12 2024-12-02T17:28:06,188 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:06,188 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.HRegion(2902): Flushing 8c1b9a0082d5e51306447fb7be8c0154 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T17:28:06,192 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/4fe19b7bf9374548a4f9659d1a001f3e is 1080, key is row0003/info:/1733160486022/Put/seqid=0 2024-12-02T17:28:06,197 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741840_1016 (size=6033) 2024-12-02T17:28:06,197 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741840_1016 (size=6033) 2024-12-02T17:28:06,197 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=13 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/4fe19b7bf9374548a4f9659d1a001f3e 2024-12-02T17:28:06,204 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/4fe19b7bf9374548a4f9659d1a001f3e as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/4fe19b7bf9374548a4f9659d1a001f3e 2024-12-02T17:28:06,209 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/4fe19b7bf9374548a4f9659d1a001f3e, entries=1, sequenceid=13, filesize=5.9 K 2024-12-02T17:28:06,210 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.HRegion(3140): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 8c1b9a0082d5e51306447fb7be8c0154 in 22ms, sequenceid=13, compaction requested=true 2024-12-02T17:28:06,210 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.HRegion(2603): Flush status journal for 8c1b9a0082d5e51306447fb7be8c0154: 2024-12-02T17:28:06,210 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:06,210 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-2 {event_type=RS_FLUSH_REGIONS, pid=12}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=12 2024-12-02T17:28:06,211 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster(4169): Remote procedure done, pid=12 2024-12-02T17:28:06,215 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=12, resume processing ppid=11 2024-12-02T17:28:06,215 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=12, ppid=11, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 178 msec 2024-12-02T17:28:06,217 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1521): Finished pid=11, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=11, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 185 msec 2024-12-02T17:28:15,715 INFO [master/3b3375d1a925:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T17:28:15,715 INFO [master/3b3375d1a925:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T17:28:16,130 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=11 2024-12-02T17:28:16,130 INFO [RPCClient-NioEventLoopGroup-4-3 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling completed 2024-12-02T17:28:16,130 DEBUG [Time-limited test {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:28:16,131 DEBUG [Time-limited test {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 18099 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:28:16,131 DEBUG [Time-limited test {}] regionserver.HStore(1541): 8c1b9a0082d5e51306447fb7be8c0154/info is initiating minor compaction (all files) 2024-12-02T17:28:16,131 INFO [Time-limited test {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:28:16,132 INFO [Time-limited test {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:16,132 INFO [Time-limited test {}] regionserver.HRegion(2416): Starting compaction of 8c1b9a0082d5e51306447fb7be8c0154/info in TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:16,132 INFO [Time-limited test {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/5ea7b4cdbd874cafb117575d7d0eb932, hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/19a7be03f1b74211ac10f6a6eb303545, hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/4fe19b7bf9374548a4f9659d1a001f3e] into tmpdir=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp, totalSize=17.7 K 2024-12-02T17:28:16,132 DEBUG [Time-limited test {}] compactions.Compactor(225): Compacting 5ea7b4cdbd874cafb117575d7d0eb932, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=5, earliestPutTs=1733160465907 2024-12-02T17:28:16,133 DEBUG [Time-limited test {}] compactions.Compactor(225): Compacting 19a7be03f1b74211ac10f6a6eb303545, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=9, earliestPutTs=1733160475941 2024-12-02T17:28:16,133 DEBUG [Time-limited test {}] compactions.Compactor(225): Compacting 4fe19b7bf9374548a4f9659d1a001f3e, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=13, earliestPutTs=1733160486022 2024-12-02T17:28:16,143 INFO [Time-limited test {}] throttle.PressureAwareThroughputController(145): 8c1b9a0082d5e51306447fb7be8c0154#info#compaction#45 average throughput is unlimited, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:28:16,144 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/56686065e4b045a599d1e0a9e918f95d is 1080, key is row0001/info:/1733160465907/Put/seqid=0 2024-12-02T17:28:16,148 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741841_1017 (size=8296) 2024-12-02T17:28:16,148 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741841_1017 (size=8296) 2024-12-02T17:28:16,153 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/56686065e4b045a599d1e0a9e918f95d as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/56686065e4b045a599d1e0a9e918f95d 2024-12-02T17:28:16,159 INFO [Time-limited test {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in 8c1b9a0082d5e51306447fb7be8c0154/info of 8c1b9a0082d5e51306447fb7be8c0154 into 56686065e4b045a599d1e0a9e918f95d(size=8.1 K), total size for store is 8.1 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:28:16,159 DEBUG [Time-limited test {}] regionserver.HRegion(2446): Compaction status journal for 8c1b9a0082d5e51306447fb7be8c0154: 2024-12-02T17:28:16,161 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C33741%2C1733160454876.1733160496161 2024-12-02T17:28:16,168 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:16,168 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:16,168 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:16,168 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:16,169 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:16,169 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160486023 with entries=4, filesize=2.45 KB; new WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160496161 2024-12-02T17:28:16,170 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741839_1015 (size=2520) 2024-12-02T17:28:16,171 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46083:46083),(127.0.0.1/127.0.0.1:35973:35973)] 2024-12-02T17:28:16,171 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741839_1015 (size=2520) 2024-12-02T17:28:16,171 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160486023 is not closed yet, will try archiving it next time 2024-12-02T17:28:16,171 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160455260 to hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/oldWALs/3b3375d1a925%2C33741%2C1733160454876.1733160455260 2024-12-02T17:28:16,172 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:28:16,173 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] procedure2.ProcedureExecutor(1139): Stored pid=13, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=13, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:28:16,174 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=13 2024-12-02T17:28:16,174 INFO [PEWorker-4 {}] procedure.FlushTableProcedure(91): pid=13, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=13, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-02T17:28:16,175 INFO [PEWorker-4 {}] procedure.FlushTableProcedure(91): pid=13, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=13, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T17:28:16,175 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=14, ppid=13, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T17:28:16,328 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=33741 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=14 2024-12-02T17:28:16,328 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:16,328 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.HRegion(2902): Flushing 8c1b9a0082d5e51306447fb7be8c0154 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T17:28:16,332 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/7a86beafb4be43a3a1065cffc2dcfaaf is 1080, key is row0000/info:/1733160496160/Put/seqid=0 2024-12-02T17:28:16,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741843_1019 (size=6033) 2024-12-02T17:28:16,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741843_1019 (size=6033) 2024-12-02T17:28:16,338 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=18 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/7a86beafb4be43a3a1065cffc2dcfaaf 2024-12-02T17:28:16,344 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/7a86beafb4be43a3a1065cffc2dcfaaf as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/7a86beafb4be43a3a1065cffc2dcfaaf 2024-12-02T17:28:16,348 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/7a86beafb4be43a3a1065cffc2dcfaaf, entries=1, sequenceid=18, filesize=5.9 K 2024-12-02T17:28:16,349 INFO [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.HRegion(3140): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 8c1b9a0082d5e51306447fb7be8c0154 in 21ms, sequenceid=18, compaction requested=false 2024-12-02T17:28:16,349 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.HRegion(2603): Flush status journal for 8c1b9a0082d5e51306447fb7be8c0154: 2024-12-02T17:28:16,349 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:16,349 DEBUG [RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0-0 {event_type=RS_FLUSH_REGIONS, pid=14}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=14 2024-12-02T17:28:16,350 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.HMaster(4169): Remote procedure done, pid=14 2024-12-02T17:28:16,353 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=14, resume processing ppid=13 2024-12-02T17:28:16,353 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=14, ppid=13, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 176 msec 2024-12-02T17:28:16,356 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=13, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=13, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 182 msec 2024-12-02T17:28:21,141 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(329): Region 8c1b9a0082d5e51306447fb7be8c0154, had cached 0 bytes from a total of 14329 2024-12-02T17:28:26,210 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46201 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=13 2024-12-02T17:28:26,210 INFO [RPCClient-NioEventLoopGroup-4-3 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling completed 2024-12-02T17:28:26,213 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C33741%2C1733160454876.1733160506213 2024-12-02T17:28:26,219 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,219 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,219 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,219 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,219 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,220 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160496161 with entries=3, filesize=1.97 KB; new WAL /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160506213 2024-12-02T17:28:26,220 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46083:46083),(127.0.0.1/127.0.0.1:35973:35973)] 2024-12-02T17:28:26,220 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160496161 is not closed yet, will try archiving it next time 2024-12-02T17:28:26,220 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/WALs/3b3375d1a925,33741,1733160454876/3b3375d1a925%2C33741%2C1733160454876.1733160486023 to hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/oldWALs/3b3375d1a925%2C33741%2C1733160454876.1733160486023 2024-12-02T17:28:26,220 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T17:28:26,221 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:28:26,221 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:28:26,221 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:28:26,221 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:28:26,221 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T17:28:26,221 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=370189627, stopped=false 2024-12-02T17:28:26,221 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=3b3375d1a925,46201,1733160454835 2024-12-02T17:28:26,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741842_1018 (size=2026) 2024-12-02T17:28:26,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741842_1018 (size=2026) 2024-12-02T17:28:26,222 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T17:28:26,223 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:28:26,223 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:26,223 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:28:26,223 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:28:26,223 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:26,223 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:28:26,223 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:28:26,223 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:28:26,224 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '3b3375d1a925,33741,1733160454876' ***** 2024-12-02T17:28:26,224 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T17:28:26,224 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T17:28:26,224 INFO [RS:0;3b3375d1a925:33741 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T17:28:26,224 INFO [RS:0;3b3375d1a925:33741 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T17:28:26,224 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(3091): Received CLOSE for 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:28:26,224 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T17:28:26,225 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(959): stopping server 3b3375d1a925,33741,1733160454876 2024-12-02T17:28:26,225 INFO [RS:0;3b3375d1a925:33741 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:28:26,225 INFO [RS:0;3b3375d1a925:33741 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;3b3375d1a925:33741. 2024-12-02T17:28:26,225 DEBUG [RS:0;3b3375d1a925:33741 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:28:26,225 DEBUG [RS:0;3b3375d1a925:33741 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:28:26,225 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 8c1b9a0082d5e51306447fb7be8c0154, disabling compactions & flushes 2024-12-02T17:28:26,225 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:26,225 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:26,225 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T17:28:26,225 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T17:28:26,225 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. after waiting 0 ms 2024-12-02T17:28:26,225 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T17:28:26,225 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:26,225 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T17:28:26,225 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2902): Flushing 8c1b9a0082d5e51306447fb7be8c0154 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T17:28:26,225 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1321): Waiting on 2 regions to close 2024-12-02T17:28:26,225 DEBUG [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740, 8c1b9a0082d5e51306447fb7be8c0154=TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.} 2024-12-02T17:28:26,225 DEBUG [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, 8c1b9a0082d5e51306447fb7be8c0154 2024-12-02T17:28:26,225 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:28:26,225 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:28:26,225 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:28:26,225 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:28:26,225 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:28:26,225 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.89 KB heapSize=3.91 KB 2024-12-02T17:28:26,229 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/8373a43112fc453da36d449c11be728a is 1080, key is row0001/info:/1733160506211/Put/seqid=0 2024-12-02T17:28:26,233 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:28:26,233 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:28:26,242 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741845_1021 (size=6033) 2024-12-02T17:28:26,242 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741845_1021 (size=6033) 2024-12-02T17:28:26,243 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=22 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/8373a43112fc453da36d449c11be728a 2024-12-02T17:28:26,249 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/.tmp/info/8373a43112fc453da36d449c11be728a as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/8373a43112fc453da36d449c11be728a 2024-12-02T17:28:26,250 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/info/bb9359fc2a47465784e8b402a5cc7829 is 227, key is TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154./info:regioninfo/1733160456154/Put/seqid=0 2024-12-02T17:28:26,257 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/8373a43112fc453da36d449c11be728a, entries=1, sequenceid=22, filesize=5.9 K 2024-12-02T17:28:26,258 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3140): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 8c1b9a0082d5e51306447fb7be8c0154 in 33ms, sequenceid=22, compaction requested=true 2024-12-02T17:28:26,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741846_1022 (size=7308) 2024-12-02T17:28:26,259 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.-1 {}] regionserver.HStore(2317): Moving the files [hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/5ea7b4cdbd874cafb117575d7d0eb932, hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/19a7be03f1b74211ac10f6a6eb303545, hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/4fe19b7bf9374548a4f9659d1a001f3e] to archive 2024-12-02T17:28:26,259 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741846_1022 (size=7308) 2024-12-02T17:28:26,259 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.65 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/info/bb9359fc2a47465784e8b402a5cc7829 2024-12-02T17:28:26,260 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T17:28:26,262 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/5ea7b4cdbd874cafb117575d7d0eb932 to hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/5ea7b4cdbd874cafb117575d7d0eb932 2024-12-02T17:28:26,263 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/19a7be03f1b74211ac10f6a6eb303545 to hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/19a7be03f1b74211ac10f6a6eb303545 2024-12-02T17:28:26,265 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/4fe19b7bf9374548a4f9659d1a001f3e to hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/info/4fe19b7bf9374548a4f9659d1a001f3e 2024-12-02T17:28:26,265 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.-1 {}] regionserver.HRegionServer(3532): Failed to report file archival(s) to Master. This will be retried. org.apache.hadoop.hbase.ipc.StoppedRpcClientException: Call to address=3b3375d1a925:46201 failed on local exception: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.hbase.ipc.IPCUtil.wrapException(IPCUtil.java:237) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:395) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:430) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:425) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:117) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.setException(Call.java:132) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:451) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:336) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:606) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$BlockingStub.reportFileArchival(RegionServerStatusProtos.java:17350) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegionServer.reportFileArchivalForQuotas(HRegionServer.java:3516) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.reportArchivedFilesForQuota(HStore.java:2412) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.removeCompactedfiles(HStore.java:2347) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.closeWithoutLock(HStore.java:738) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.close(HStore.java:804) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion$2.call(HRegion.java:1912) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion$2.call(HRegion.java:1909) ~[classes/:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) ~[?:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at org.apache.hadoop.hbase.ipc.AbstractRpcClient.getConnection(AbstractRpcClient.java:366) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:448) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] ... 16 more 2024-12-02T17:28:26,266 WARN [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154.-1 {}] regionserver.HStore(2414): Failed to report archival of files: [5ea7b4cdbd874cafb117575d7d0eb932=6033, 19a7be03f1b74211ac10f6a6eb303545=6033, 4fe19b7bf9374548a4f9659d1a001f3e=6033] 2024-12-02T17:28:26,270 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/8c1b9a0082d5e51306447fb7be8c0154/recovered.edits/25.seqid, newMaxSeqId=25, maxSeqId=1 2024-12-02T17:28:26,271 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:26,271 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 8c1b9a0082d5e51306447fb7be8c0154: Waiting for close lock at 1733160506225Running coprocessor pre-close hooks at 1733160506225Disabling compacts and flushes for region at 1733160506225Disabling writes for close at 1733160506225Obtaining lock to block concurrent updates at 1733160506225Preparing flush snapshotting stores in 8c1b9a0082d5e51306447fb7be8c0154 at 1733160506225Finished memstore snapshotting TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154., syncing WAL and waiting on mvcc, flushsize=dataSize=1076, getHeapSize=1392, getOffHeapSize=0, getCellsCount=1 at 1733160506225Flushing stores of TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. at 1733160506226 (+1 ms)Flushing 8c1b9a0082d5e51306447fb7be8c0154/info: creating writer at 1733160506226Flushing 8c1b9a0082d5e51306447fb7be8c0154/info: appending metadata at 1733160506228 (+2 ms)Flushing 8c1b9a0082d5e51306447fb7be8c0154/info: closing flushed file at 1733160506229 (+1 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1e5d7bfe: reopening flushed file at 1733160506248 (+19 ms)Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 8c1b9a0082d5e51306447fb7be8c0154 in 33ms, sequenceid=22, compaction requested=true at 1733160506258 (+10 ms)Writing region close event to WAL at 1733160506266 (+8 ms)Running coprocessor post-close hooks at 1733160506271 (+5 ms)Closed at 1733160506271 2024-12-02T17:28:26,271 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733160455804.8c1b9a0082d5e51306447fb7be8c0154. 2024-12-02T17:28:26,281 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/ns/8d07d0ce19db4a79bb067b28df536cbd is 43, key is default/ns:d/1733160455697/Put/seqid=0 2024-12-02T17:28:26,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741847_1023 (size=5153) 2024-12-02T17:28:26,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741847_1023 (size=5153) 2024-12-02T17:28:26,286 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/ns/8d07d0ce19db4a79bb067b28df536cbd 2024-12-02T17:28:26,304 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/table/80cf13f19c5d4520a47c392a57c650e1 is 89, key is TestLogRolling-testCompactionRecordDoesntBlockRolling/table:state/1733160456164/Put/seqid=0 2024-12-02T17:28:26,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741848_1024 (size=5508) 2024-12-02T17:28:26,309 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741848_1024 (size=5508) 2024-12-02T17:28:26,309 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=170 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/table/80cf13f19c5d4520a47c392a57c650e1 2024-12-02T17:28:26,314 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/info/bb9359fc2a47465784e8b402a5cc7829 as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/info/bb9359fc2a47465784e8b402a5cc7829 2024-12-02T17:28:26,320 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/info/bb9359fc2a47465784e8b402a5cc7829, entries=10, sequenceid=11, filesize=7.1 K 2024-12-02T17:28:26,321 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/ns/8d07d0ce19db4a79bb067b28df536cbd as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/ns/8d07d0ce19db4a79bb067b28df536cbd 2024-12-02T17:28:26,326 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/ns/8d07d0ce19db4a79bb067b28df536cbd, entries=2, sequenceid=11, filesize=5.0 K 2024-12-02T17:28:26,326 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/.tmp/table/80cf13f19c5d4520a47c392a57c650e1 as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/table/80cf13f19c5d4520a47c392a57c650e1 2024-12-02T17:28:26,330 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/table/80cf13f19c5d4520a47c392a57c650e1, entries=2, sequenceid=11, filesize=5.4 K 2024-12-02T17:28:26,331 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.89 KB/1932, heapSize ~3.61 KB/3696, currentSize=0 B/0 for 1588230740 in 106ms, sequenceid=11, compaction requested=false 2024-12-02T17:28:26,335 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-02T17:28:26,336 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:28:26,336 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:28:26,336 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160506225Running coprocessor pre-close hooks at 1733160506225Disabling compacts and flushes for region at 1733160506225Disabling writes for close at 1733160506225Obtaining lock to block concurrent updates at 1733160506225Preparing flush snapshotting stores in 1588230740 at 1733160506225Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1932, getHeapSize=3936, getOffHeapSize=0, getCellsCount=14 at 1733160506226 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733160506226Flushing 1588230740/info: creating writer at 1733160506226Flushing 1588230740/info: appending metadata at 1733160506250 (+24 ms)Flushing 1588230740/info: closing flushed file at 1733160506250Flushing 1588230740/ns: creating writer at 1733160506265 (+15 ms)Flushing 1588230740/ns: appending metadata at 1733160506280 (+15 ms)Flushing 1588230740/ns: closing flushed file at 1733160506280Flushing 1588230740/table: creating writer at 1733160506291 (+11 ms)Flushing 1588230740/table: appending metadata at 1733160506304 (+13 ms)Flushing 1588230740/table: closing flushed file at 1733160506304Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5b058d3a: reopening flushed file at 1733160506314 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2d2a71b5: reopening flushed file at 1733160506320 (+6 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3c0ff187: reopening flushed file at 1733160506326 (+6 ms)Finished flush of dataSize ~1.89 KB/1932, heapSize ~3.61 KB/3696, currentSize=0 B/0 for 1588230740 in 106ms, sequenceid=11, compaction requested=false at 1733160506331 (+5 ms)Writing region close event to WAL at 1733160506332 (+1 ms)Running coprocessor post-close hooks at 1733160506336 (+4 ms)Closed at 1733160506336 2024-12-02T17:28:26,336 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T17:28:26,425 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(976): stopping server 3b3375d1a925,33741,1733160454876; all regions closed. 2024-12-02T17:28:26,426 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,426 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,426 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,426 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,426 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741834_1010 (size=3306) 2024-12-02T17:28:26,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741834_1010 (size=3306) 2024-12-02T17:28:26,430 DEBUG [RS:0;3b3375d1a925:33741 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/oldWALs 2024-12-02T17:28:26,430 INFO [RS:0;3b3375d1a925:33741 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C33741%2C1733160454876.meta:.meta(num 1733160455661) 2024-12-02T17:28:26,430 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,430 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,430 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,431 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,431 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741844_1020 (size=1252) 2024-12-02T17:28:26,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741844_1020 (size=1252) 2024-12-02T17:28:26,435 DEBUG [RS:0;3b3375d1a925:33741 {}] wal.AbstractFSWAL(1256): Moved 2 WAL file(s) to /user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/oldWALs 2024-12-02T17:28:26,435 INFO [RS:0;3b3375d1a925:33741 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C33741%2C1733160454876:(num 1733160506213) 2024-12-02T17:28:26,435 DEBUG [RS:0;3b3375d1a925:33741 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:28:26,435 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:28:26,435 INFO [RS:0;3b3375d1a925:33741 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:28:26,435 INFO [RS:0;3b3375d1a925:33741 {}] hbase.ChoreService(370): Chore service for: regionserver/3b3375d1a925:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-02T17:28:26,435 INFO [RS:0;3b3375d1a925:33741 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:28:26,435 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:28:26,436 INFO [RS:0;3b3375d1a925:33741 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:33741 2024-12-02T17:28:26,438 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/3b3375d1a925,33741,1733160454876 2024-12-02T17:28:26,438 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:28:26,438 INFO [RS:0;3b3375d1a925:33741 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:28:26,438 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [3b3375d1a925,33741,1733160454876] 2024-12-02T17:28:26,440 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/3b3375d1a925,33741,1733160454876 already deleted, retry=false 2024-12-02T17:28:26,440 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 3b3375d1a925,33741,1733160454876 expired; onlineServers=0 2024-12-02T17:28:26,440 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '3b3375d1a925,46201,1733160454835' ***** 2024-12-02T17:28:26,440 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T17:28:26,440 INFO [M:0;3b3375d1a925:46201 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:28:26,440 INFO [M:0;3b3375d1a925:46201 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:28:26,440 DEBUG [M:0;3b3375d1a925:46201 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T17:28:26,440 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T17:28:26,440 DEBUG [M:0;3b3375d1a925:46201 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T17:28:26,440 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160455054 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160455054,5,FailOnTimeoutGroup] 2024-12-02T17:28:26,440 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160455054 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160455054,5,FailOnTimeoutGroup] 2024-12-02T17:28:26,440 INFO [M:0;3b3375d1a925:46201 {}] hbase.ChoreService(370): Chore service for: master/3b3375d1a925:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T17:28:26,440 INFO [M:0;3b3375d1a925:46201 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:28:26,441 DEBUG [M:0;3b3375d1a925:46201 {}] master.HMaster(1795): Stopping service threads 2024-12-02T17:28:26,441 INFO [M:0;3b3375d1a925:46201 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T17:28:26,441 INFO [M:0;3b3375d1a925:46201 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:28:26,441 INFO [M:0;3b3375d1a925:46201 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T17:28:26,441 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T17:28:26,442 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T17:28:26,442 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:26,442 DEBUG [M:0;3b3375d1a925:46201 {}] zookeeper.ZKUtil(347): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T17:28:26,442 WARN [M:0;3b3375d1a925:46201 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T17:28:26,442 INFO [M:0;3b3375d1a925:46201 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/.lastflushedseqids 2024-12-02T17:28:26,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741849_1025 (size=130) 2024-12-02T17:28:26,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741849_1025 (size=130) 2024-12-02T17:28:26,448 INFO [M:0;3b3375d1a925:46201 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T17:28:26,448 INFO [M:0;3b3375d1a925:46201 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T17:28:26,448 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:28:26,448 INFO [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:28:26,448 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:28:26,448 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:28:26,448 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:28:26,448 INFO [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=43.53 KB heapSize=54.89 KB 2024-12-02T17:28:26,470 DEBUG [M:0;3b3375d1a925:46201 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/002c290470034008a50ebb4f34ceedec is 82, key is hbase:meta,,1/info:regioninfo/1733160455683/Put/seqid=0 2024-12-02T17:28:26,474 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741850_1026 (size=5672) 2024-12-02T17:28:26,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741850_1026 (size=5672) 2024-12-02T17:28:26,475 INFO [M:0;3b3375d1a925:46201 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=121 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/002c290470034008a50ebb4f34ceedec 2024-12-02T17:28:26,493 DEBUG [M:0;3b3375d1a925:46201 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6950a178824042db834eaff3998c5cfb is 797, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733160456168/Put/seqid=0 2024-12-02T17:28:26,497 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741851_1027 (size=7816) 2024-12-02T17:28:26,497 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741851_1027 (size=7816) 2024-12-02T17:28:26,498 INFO [M:0;3b3375d1a925:46201 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=42.92 KB at sequenceid=121 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6950a178824042db834eaff3998c5cfb 2024-12-02T17:28:26,502 INFO [M:0;3b3375d1a925:46201 {}] regionserver.StoreFileReader(518): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 6950a178824042db834eaff3998c5cfb 2024-12-02T17:28:26,522 DEBUG [M:0;3b3375d1a925:46201 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/93b6f76c1d1349849a2ed9752e3a97b8 is 69, key is 3b3375d1a925,33741,1733160454876/rs:state/1733160455112/Put/seqid=0 2024-12-02T17:28:26,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741852_1028 (size=5156) 2024-12-02T17:28:26,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741852_1028 (size=5156) 2024-12-02T17:28:26,526 INFO [M:0;3b3375d1a925:46201 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=121 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/93b6f76c1d1349849a2ed9752e3a97b8 2024-12-02T17:28:26,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:28:26,539 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33741-0x1009c07d5450001, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:28:26,539 INFO [RS:0;3b3375d1a925:33741 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:28:26,539 INFO [RS:0;3b3375d1a925:33741 {}] regionserver.HRegionServer(1031): Exiting; stopping=3b3375d1a925,33741,1733160454876; zookeeper connection closed. 2024-12-02T17:28:26,539 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5286f706 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5286f706 2024-12-02T17:28:26,540 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T17:28:26,543 DEBUG [M:0;3b3375d1a925:46201 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/1865320964dc449789aa021e6f7cd5c2 is 52, key is load_balancer_on/state:d/1733160455801/Put/seqid=0 2024-12-02T17:28:26,547 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741853_1029 (size=5056) 2024-12-02T17:28:26,548 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741853_1029 (size=5056) 2024-12-02T17:28:26,548 INFO [M:0;3b3375d1a925:46201 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=121 (bloomFilter=true), to=hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/1865320964dc449789aa021e6f7cd5c2 2024-12-02T17:28:26,553 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/002c290470034008a50ebb4f34ceedec as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/002c290470034008a50ebb4f34ceedec 2024-12-02T17:28:26,557 INFO [M:0;3b3375d1a925:46201 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/002c290470034008a50ebb4f34ceedec, entries=8, sequenceid=121, filesize=5.5 K 2024-12-02T17:28:26,558 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6950a178824042db834eaff3998c5cfb as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/6950a178824042db834eaff3998c5cfb 2024-12-02T17:28:26,561 INFO [M:0;3b3375d1a925:46201 {}] regionserver.StoreFileReader(518): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 6950a178824042db834eaff3998c5cfb 2024-12-02T17:28:26,562 INFO [M:0;3b3375d1a925:46201 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/6950a178824042db834eaff3998c5cfb, entries=14, sequenceid=121, filesize=7.6 K 2024-12-02T17:28:26,562 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/93b6f76c1d1349849a2ed9752e3a97b8 as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/93b6f76c1d1349849a2ed9752e3a97b8 2024-12-02T17:28:26,566 INFO [M:0;3b3375d1a925:46201 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/93b6f76c1d1349849a2ed9752e3a97b8, entries=1, sequenceid=121, filesize=5.0 K 2024-12-02T17:28:26,567 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/1865320964dc449789aa021e6f7cd5c2 as hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/1865320964dc449789aa021e6f7cd5c2 2024-12-02T17:28:26,570 INFO [M:0;3b3375d1a925:46201 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42485/user/jenkins/test-data/c4e6aa00-ddc5-4731-38eb-faa409c38c32/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/1865320964dc449789aa021e6f7cd5c2, entries=1, sequenceid=121, filesize=4.9 K 2024-12-02T17:28:26,571 INFO [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(3140): Finished flush of dataSize ~43.53 KB/44572, heapSize ~54.83 KB/56144, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 123ms, sequenceid=121, compaction requested=false 2024-12-02T17:28:26,573 INFO [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:28:26,573 DEBUG [M:0;3b3375d1a925:46201 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160506448Disabling compacts and flushes for region at 1733160506448Disabling writes for close at 1733160506448Obtaining lock to block concurrent updates at 1733160506448Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733160506448Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=44572, getHeapSize=56144, getOffHeapSize=0, getCellsCount=140 at 1733160506449 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733160506449Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733160506449Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733160506470 (+21 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733160506470Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733160506479 (+9 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733160506493 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733160506493Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733160506502 (+9 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733160506521 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733160506521Flushing 1595e783b53d99cd5eef43b6debb2682/state: creating writer at 1733160506530 (+9 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: appending metadata at 1733160506543 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: closing flushed file at 1733160506543Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@b26ab2f: reopening flushed file at 1733160506552 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6fb4b9f5: reopening flushed file at 1733160506557 (+5 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@51524f7c: reopening flushed file at 1733160506562 (+5 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@42c0f71f: reopening flushed file at 1733160506566 (+4 ms)Finished flush of dataSize ~43.53 KB/44572, heapSize ~54.83 KB/56144, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 123ms, sequenceid=121, compaction requested=false at 1733160506571 (+5 ms)Writing region close event to WAL at 1733160506573 (+2 ms)Closed at 1733160506573 2024-12-02T17:28:26,573 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,573 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,573 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,574 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,574 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:28:26,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39997 is added to blk_1073741830_1006 (size=52969) 2024-12-02T17:28:26,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45461 is added to blk_1073741830_1006 (size=52969) 2024-12-02T17:28:26,576 INFO [M:0;3b3375d1a925:46201 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T17:28:26,576 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:28:26,576 INFO [M:0;3b3375d1a925:46201 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:46201 2024-12-02T17:28:26,576 INFO [M:0;3b3375d1a925:46201 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:28:26,678 INFO [M:0;3b3375d1a925:46201 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:28:26,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:28:26,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46201-0x1009c07d5450000, quorum=127.0.0.1:63626, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:28:26,680 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@191911fe{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:28:26,681 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@30d9f702{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:28:26,681 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:28:26,681 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4dfe8781{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:28:26,681 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6df2cf02{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.log.dir/,STOPPED} 2024-12-02T17:28:26,683 WARN [BP-1826006310-172.17.0.3-1733160454087 heartbeating to localhost/127.0.0.1:42485 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:28:26,683 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:28:26,683 WARN [BP-1826006310-172.17.0.3-1733160454087 heartbeating to localhost/127.0.0.1:42485 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1826006310-172.17.0.3-1733160454087 (Datanode Uuid 783409ae-4421-47a1-9609-9b02818e9a25) service to localhost/127.0.0.1:42485 2024-12-02T17:28:26,683 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:28:26,683 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/data/data3/current/BP-1826006310-172.17.0.3-1733160454087 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:28:26,684 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/data/data4/current/BP-1826006310-172.17.0.3-1733160454087 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:28:26,684 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:28:26,686 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1af676f5{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:28:26,686 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@9cd716d{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:28:26,686 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:28:26,686 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5718d675{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:28:26,686 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@25c02940{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.log.dir/,STOPPED} 2024-12-02T17:28:26,688 WARN [BP-1826006310-172.17.0.3-1733160454087 heartbeating to localhost/127.0.0.1:42485 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:28:26,688 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:28:26,688 WARN [BP-1826006310-172.17.0.3-1733160454087 heartbeating to localhost/127.0.0.1:42485 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1826006310-172.17.0.3-1733160454087 (Datanode Uuid e4714d92-0927-492d-8fb0-ba7f5f536e47) service to localhost/127.0.0.1:42485 2024-12-02T17:28:26,688 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:28:26,689 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/data/data1/current/BP-1826006310-172.17.0.3-1733160454087 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:28:26,689 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/cluster_374b5d0c-9fcf-dc6b-6a0b-f28bd561449c/data/data2/current/BP-1826006310-172.17.0.3-1733160454087 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:28:26,689 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:28:26,695 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@47df503f{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:28:26,696 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5f3519da{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:28:26,696 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:28:26,696 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4a31a089{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:28:26,696 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@15cd018{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.log.dir/,STOPPED} 2024-12-02T17:28:26,702 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T17:28:26,718 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T17:28:26,727 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testCompactionRecordDoesntBlockRolling Thread=203 (was 177) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42485 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-35-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42485 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42485 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-35-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:42485 from jenkins.hfs.5 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: HMaster-EventLoopGroup-12-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-12-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-35-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-36-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:42485 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-37-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-34-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-34-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-12-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-37-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:42485 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-16 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-13-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-13-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-37-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: regionserver/3b3375d1a925:0.leaseChecker java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.regionserver.LeaseManager.run(LeaseManager.java:82) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:42485 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-13-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-36-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.5@localhost:42485 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-34-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-36-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=483 (was 457) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=133 (was 178), ProcessCount=11 (was 11), AvailableMemoryMB=3184 (was 2147) - AvailableMemoryMB LEAK? - 2024-12-02T17:28:26,735 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRolling Thread=203, OpenFileDescriptor=483, MaxFileDescriptor=1048576, SystemLoadAverage=133, ProcessCount=11, AvailableMemoryMB=3184 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.log.dir so I do NOT create it in target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/3ee61cd6-fa1e-01e5-bc38-20c095b111bd/hadoop.tmp.dir so I do NOT create it in target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f, deleteOnExit=true 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/test.cache.data in system properties and HBase conf 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.log.dir in system properties and HBase conf 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T17:28:26,736 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T17:28:26,736 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/nfs.dump.dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/java.io.tmpdir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T17:28:26,737 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T17:28:26,750 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:28:26,801 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:28:26,804 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:28:26,805 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:28:26,805 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:28:26,806 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:28:26,806 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:28:26,806 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2ffc76d1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:28:26,807 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@273a6f23{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:28:26,921 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@40a5a9d2{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/java.io.tmpdir/jetty-localhost-43137-hadoop-hdfs-3_4_1-tests_jar-_-any-1889439090693204992/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:28:26,922 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@655ba446{HTTP/1.1, (http/1.1)}{localhost:43137} 2024-12-02T17:28:26,922 INFO [Time-limited test {}] server.Server(415): Started @234594ms 2024-12-02T17:28:26,934 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:28:26,985 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:28:26,987 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:28:26,988 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:28:26,988 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:28:26,988 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:28:26,989 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3aa952ac{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:28:26,989 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3aca12f3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:28:27,103 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@e1b4695{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/java.io.tmpdir/jetty-localhost-33723-hadoop-hdfs-3_4_1-tests_jar-_-any-6137469413823976725/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:28:27,103 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1520cb76{HTTP/1.1, (http/1.1)}{localhost:33723} 2024-12-02T17:28:27,103 INFO [Time-limited test {}] server.Server(415): Started @234775ms 2024-12-02T17:28:27,104 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:28:27,129 INFO [regionserver/3b3375d1a925:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:28:27,133 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:28:27,135 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:28:27,136 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:28:27,136 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:28:27,136 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:28:27,136 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5fe51576{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:28:27,137 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5b6783f5{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:28:27,180 WARN [Thread-1937 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/data/data1/current/BP-436516782-172.17.0.3-1733160506756/current, will proceed with Du for space computation calculation, 2024-12-02T17:28:27,180 WARN [Thread-1938 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/data/data2/current/BP-436516782-172.17.0.3-1733160506756/current, will proceed with Du for space computation calculation, 2024-12-02T17:28:27,199 WARN [Thread-1916 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:28:27,201 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe9969ee9cf5b7a6b with lease ID 0x2291171faff86843: Processing first storage report for DS-53389ce6-c317-479b-818c-87972c9a3b8a from datanode DatanodeRegistration(127.0.0.1:42947, datanodeUuid=ee7e4b05-cc51-4ee8-840f-45b80a5cd1c2, infoPort=40389, infoSecurePort=0, ipcPort=37011, storageInfo=lv=-57;cid=testClusterID;nsid=1691333734;c=1733160506756) 2024-12-02T17:28:27,201 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe9969ee9cf5b7a6b with lease ID 0x2291171faff86843: from storage DS-53389ce6-c317-479b-818c-87972c9a3b8a node DatanodeRegistration(127.0.0.1:42947, datanodeUuid=ee7e4b05-cc51-4ee8-840f-45b80a5cd1c2, infoPort=40389, infoSecurePort=0, ipcPort=37011, storageInfo=lv=-57;cid=testClusterID;nsid=1691333734;c=1733160506756), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:28:27,201 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe9969ee9cf5b7a6b with lease ID 0x2291171faff86843: Processing first storage report for DS-f3f3b62e-6b3c-42cf-a217-a5f594bb6c20 from datanode DatanodeRegistration(127.0.0.1:42947, datanodeUuid=ee7e4b05-cc51-4ee8-840f-45b80a5cd1c2, infoPort=40389, infoSecurePort=0, ipcPort=37011, storageInfo=lv=-57;cid=testClusterID;nsid=1691333734;c=1733160506756) 2024-12-02T17:28:27,201 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe9969ee9cf5b7a6b with lease ID 0x2291171faff86843: from storage DS-f3f3b62e-6b3c-42cf-a217-a5f594bb6c20 node DatanodeRegistration(127.0.0.1:42947, datanodeUuid=ee7e4b05-cc51-4ee8-840f-45b80a5cd1c2, infoPort=40389, infoSecurePort=0, ipcPort=37011, storageInfo=lv=-57;cid=testClusterID;nsid=1691333734;c=1733160506756), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:28:27,251 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@14646d9{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/java.io.tmpdir/jetty-localhost-44093-hadoop-hdfs-3_4_1-tests_jar-_-any-6830957900296537402/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:28:27,251 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1d428ad5{HTTP/1.1, (http/1.1)}{localhost:44093} 2024-12-02T17:28:27,251 INFO [Time-limited test {}] server.Server(415): Started @234924ms 2024-12-02T17:28:27,252 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:28:27,349 WARN [Thread-1964 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/data/data4/current/BP-436516782-172.17.0.3-1733160506756/current, will proceed with Du for space computation calculation, 2024-12-02T17:28:27,349 WARN [Thread-1963 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/data/data3/current/BP-436516782-172.17.0.3-1733160506756/current, will proceed with Du for space computation calculation, 2024-12-02T17:28:27,365 WARN [Thread-1952 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:28:27,367 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x1d4a14bceec60390 with lease ID 0x2291171faff86844: Processing first storage report for DS-458c0ad7-b67c-4dcc-bab9-c45a4ac3adff from datanode DatanodeRegistration(127.0.0.1:45313, datanodeUuid=d4cdc363-0de9-4f5d-9992-b88cee95e7a0, infoPort=35929, infoSecurePort=0, ipcPort=37103, storageInfo=lv=-57;cid=testClusterID;nsid=1691333734;c=1733160506756) 2024-12-02T17:28:27,367 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x1d4a14bceec60390 with lease ID 0x2291171faff86844: from storage DS-458c0ad7-b67c-4dcc-bab9-c45a4ac3adff node DatanodeRegistration(127.0.0.1:45313, datanodeUuid=d4cdc363-0de9-4f5d-9992-b88cee95e7a0, infoPort=35929, infoSecurePort=0, ipcPort=37103, storageInfo=lv=-57;cid=testClusterID;nsid=1691333734;c=1733160506756), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:28:27,367 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x1d4a14bceec60390 with lease ID 0x2291171faff86844: Processing first storage report for DS-ad206a0a-536a-416f-bfb7-bdbf9f42a38e from datanode DatanodeRegistration(127.0.0.1:45313, datanodeUuid=d4cdc363-0de9-4f5d-9992-b88cee95e7a0, infoPort=35929, infoSecurePort=0, ipcPort=37103, storageInfo=lv=-57;cid=testClusterID;nsid=1691333734;c=1733160506756) 2024-12-02T17:28:27,367 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x1d4a14bceec60390 with lease ID 0x2291171faff86844: from storage DS-ad206a0a-536a-416f-bfb7-bdbf9f42a38e node DatanodeRegistration(127.0.0.1:45313, datanodeUuid=d4cdc363-0de9-4f5d-9992-b88cee95e7a0, infoPort=35929, infoSecurePort=0, ipcPort=37103, storageInfo=lv=-57;cid=testClusterID;nsid=1691333734;c=1733160506756), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:28:27,393 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20 2024-12-02T17:28:27,396 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/zookeeper_0, clientPort=53832, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T17:28:27,396 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=53832 2024-12-02T17:28:27,397 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:28:27,398 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:28:27,409 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:28:27,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:28:27,411 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec with version=8 2024-12-02T17:28:27,411 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase-staging 2024-12-02T17:28:27,413 INFO [Time-limited test {}] client.ConnectionUtils(128): master/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:28:27,413 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:28:27,413 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:28:27,413 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:28:27,413 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:28:27,413 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:28:27,413 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T17:28:27,413 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:28:27,414 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:39831 2024-12-02T17:28:27,415 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:39831 connecting to ZooKeeper ensemble=127.0.0.1:53832 2024-12-02T17:28:27,420 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:398310x0, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:28:27,420 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:39831-0x1009c08a2a20000 connected 2024-12-02T17:28:27,433 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:28:27,434 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:28:27,436 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:28:27,436 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec, hbase.cluster.distributed=false 2024-12-02T17:28:27,438 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:28:27,438 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39831 2024-12-02T17:28:27,438 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39831 2024-12-02T17:28:27,438 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39831 2024-12-02T17:28:27,439 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39831 2024-12-02T17:28:27,439 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39831 2024-12-02T17:28:27,453 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:28:27,453 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:28:27,453 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:28:27,454 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:28:27,454 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:28:27,454 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:28:27,454 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T17:28:27,454 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:28:27,454 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:40037 2024-12-02T17:28:27,456 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:40037 connecting to ZooKeeper ensemble=127.0.0.1:53832 2024-12-02T17:28:27,456 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:28:27,457 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:28:27,461 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:400370x0, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:28:27,462 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:400370x0, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:28:27,462 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:40037-0x1009c08a2a20001 connected 2024-12-02T17:28:27,462 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T17:28:27,462 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T17:28:27,463 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T17:28:27,464 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:28:27,464 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40037 2024-12-02T17:28:27,464 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40037 2024-12-02T17:28:27,464 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40037 2024-12-02T17:28:27,465 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40037 2024-12-02T17:28:27,465 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40037 2024-12-02T17:28:27,476 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;3b3375d1a925:39831 2024-12-02T17:28:27,476 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/3b3375d1a925,39831,1733160507412 2024-12-02T17:28:27,478 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:28:27,478 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:28:27,478 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/3b3375d1a925,39831,1733160507412 2024-12-02T17:28:27,480 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T17:28:27,480 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,480 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,480 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T17:28:27,480 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/3b3375d1a925,39831,1733160507412 from backup master directory 2024-12-02T17:28:27,482 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/3b3375d1a925,39831,1733160507412 2024-12-02T17:28:27,482 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:28:27,482 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:28:27,482 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:28:27,482 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=3b3375d1a925,39831,1733160507412 2024-12-02T17:28:27,485 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/hbase.id] with ID: f03793dc-8a4a-47ee-9dd7-4c3d64891b97 2024-12-02T17:28:27,485 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/.tmp/hbase.id 2024-12-02T17:28:27,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:28:27,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:28:27,491 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/.tmp/hbase.id]:[hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/hbase.id] 2024-12-02T17:28:27,501 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:28:27,501 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T17:28:27,503 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 1ms. 2024-12-02T17:28:27,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:28:27,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:28:27,510 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:28:27,510 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T17:28:27,511 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:28:27,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:28:27,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:28:27,517 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store 2024-12-02T17:28:27,522 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:28:27,522 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:28:27,523 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:28:27,523 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:28:27,523 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:28:27,523 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:28:27,523 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:28:27,523 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:28:27,523 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:28:27,523 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160507523Disabling compacts and flushes for region at 1733160507523Disabling writes for close at 1733160507523Writing region close event to WAL at 1733160507523Closed at 1733160507523 2024-12-02T17:28:27,523 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/.initializing 2024-12-02T17:28:27,524 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/WALs/3b3375d1a925,39831,1733160507412 2024-12-02T17:28:27,526 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C39831%2C1733160507412, suffix=, logDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/WALs/3b3375d1a925,39831,1733160507412, archiveDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/oldWALs, maxLogs=10 2024-12-02T17:28:27,526 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C39831%2C1733160507412.1733160507526 2024-12-02T17:28:27,530 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/WALs/3b3375d1a925,39831,1733160507412/3b3375d1a925%2C39831%2C1733160507412.1733160507526 2024-12-02T17:28:27,531 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35929:35929),(127.0.0.1/127.0.0.1:40389:40389)] 2024-12-02T17:28:27,531 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:28:27,531 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:28:27,531 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,531 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,532 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,533 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T17:28:27,534 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,534 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:27,534 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,535 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T17:28:27,535 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,535 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:28:27,535 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,536 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T17:28:27,536 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,537 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:28:27,537 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,538 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T17:28:27,538 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,538 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:28:27,538 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,539 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,539 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,540 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,540 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,541 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T17:28:27,541 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:28:27,543 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:28:27,543 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=801167, jitterRate=0.018737390637397766}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T17:28:27,544 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733160507531Initializing all the Stores at 1733160507532 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160507532Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160507532Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160507532Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160507532Cleaning up temporary data from old regions at 1733160507540 (+8 ms)Region opened successfully at 1733160507544 (+4 ms) 2024-12-02T17:28:27,545 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T17:28:27,547 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@74b33613, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:28:27,548 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T17:28:27,548 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T17:28:27,548 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T17:28:27,548 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T17:28:27,549 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T17:28:27,549 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-02T17:28:27,549 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T17:28:27,551 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T17:28:27,552 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T17:28:27,553 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T17:28:27,553 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T17:28:27,554 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T17:28:27,555 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T17:28:27,556 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T17:28:27,556 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T17:28:27,557 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T17:28:27,558 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T17:28:27,559 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T17:28:27,561 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T17:28:27,562 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T17:28:27,564 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:28:27,564 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:28:27,564 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,564 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,564 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=3b3375d1a925,39831,1733160507412, sessionid=0x1009c08a2a20000, setting cluster-up flag (Was=false) 2024-12-02T17:28:27,567 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,567 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,571 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T17:28:27,572 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,39831,1733160507412 2024-12-02T17:28:27,575 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,575 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:27,580 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T17:28:27,581 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,39831,1733160507412 2024-12-02T17:28:27,582 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T17:28:27,584 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T17:28:27,584 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T17:28:27,584 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T17:28:27,584 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 3b3375d1a925,39831,1733160507412 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T17:28:27,585 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:28:27,586 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:28:27,586 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:28:27,586 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:28:27,586 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/3b3375d1a925:0, corePoolSize=10, maxPoolSize=10 2024-12-02T17:28:27,586 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,586 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:28:27,586 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,588 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:28:27,588 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T17:28:27,589 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733160537589 2024-12-02T17:28:27,589 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T17:28:27,589 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,589 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T17:28:27,589 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T17:28:27,589 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T17:28:27,589 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T17:28:27,589 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T17:28:27,589 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T17:28:27,589 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,592 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T17:28:27,592 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T17:28:27,592 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T17:28:27,592 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T17:28:27,592 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T17:28:27,592 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160507592,5,FailOnTimeoutGroup] 2024-12-02T17:28:27,593 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160507593,5,FailOnTimeoutGroup] 2024-12-02T17:28:27,593 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,593 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T17:28:27,593 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,593 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:28:27,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:28:27,597 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T17:28:27,597 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec 2024-12-02T17:28:27,602 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:28:27,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:28:27,603 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:28:27,604 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:28:27,605 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:28:27,605 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,606 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:27,606 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:28:27,607 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:28:27,607 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,607 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:27,607 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:28:27,608 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:28:27,608 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,608 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:27,609 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:28:27,609 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:28:27,609 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:27,610 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:27,610 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:28:27,610 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740 2024-12-02T17:28:27,611 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740 2024-12-02T17:28:27,612 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:28:27,612 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:28:27,612 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:28:27,613 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:28:27,615 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:28:27,615 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=810943, jitterRate=0.031168147921562195}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:28:27,615 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733160507603Initializing all the Stores at 1733160507604 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160507604Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160507604Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160507604Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160507604Cleaning up temporary data from old regions at 1733160507612 (+8 ms)Region opened successfully at 1733160507615 (+3 ms) 2024-12-02T17:28:27,616 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:28:27,616 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:28:27,616 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:28:27,616 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:28:27,616 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:28:27,616 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:28:27,616 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160507615Disabling compacts and flushes for region at 1733160507615Disabling writes for close at 1733160507616 (+1 ms)Writing region close event to WAL at 1733160507616Closed at 1733160507616 2024-12-02T17:28:27,617 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:28:27,617 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T17:28:27,617 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T17:28:27,618 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:28:27,619 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T17:28:27,666 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(746): ClusterId : f03793dc-8a4a-47ee-9dd7-4c3d64891b97 2024-12-02T17:28:27,666 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T17:28:27,669 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T17:28:27,669 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T17:28:27,671 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T17:28:27,671 DEBUG [RS:0;3b3375d1a925:40037 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3ae62aee, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:28:27,682 DEBUG [RS:0;3b3375d1a925:40037 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;3b3375d1a925:40037 2024-12-02T17:28:27,682 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T17:28:27,682 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T17:28:27,682 DEBUG [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T17:28:27,683 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(2659): reportForDuty to master=3b3375d1a925,39831,1733160507412 with port=40037, startcode=1733160507453 2024-12-02T17:28:27,683 DEBUG [RS:0;3b3375d1a925:40037 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T17:28:27,685 INFO [HMaster-EventLoopGroup-14-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49637, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.6 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T17:28:27,686 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39831 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 3b3375d1a925,40037,1733160507453 2024-12-02T17:28:27,686 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39831 {}] master.ServerManager(517): Registering regionserver=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:27,688 DEBUG [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec 2024-12-02T17:28:27,688 DEBUG [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42821 2024-12-02T17:28:27,688 DEBUG [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T17:28:27,689 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:28:27,690 DEBUG [RS:0;3b3375d1a925:40037 {}] zookeeper.ZKUtil(111): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/3b3375d1a925,40037,1733160507453 2024-12-02T17:28:27,690 WARN [RS:0;3b3375d1a925:40037 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:28:27,690 INFO [RS:0;3b3375d1a925:40037 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:28:27,690 DEBUG [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453 2024-12-02T17:28:27,690 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [3b3375d1a925,40037,1733160507453] 2024-12-02T17:28:27,693 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T17:28:27,695 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T17:28:27,695 INFO [RS:0;3b3375d1a925:40037 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:28:27,695 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,696 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T17:28:27,697 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T17:28:27,697 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:28:27,697 DEBUG [RS:0;3b3375d1a925:40037 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:28:27,698 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,698 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,698 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,698 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,698 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,698 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,40037,1733160507453-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:28:27,712 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T17:28:27,712 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,40037,1733160507453-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,712 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,712 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.Replication(171): 3b3375d1a925,40037,1733160507453 started 2024-12-02T17:28:27,726 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:27,726 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1482): Serving as 3b3375d1a925,40037,1733160507453, RpcServer on 3b3375d1a925/172.17.0.3:40037, sessionid=0x1009c08a2a20001 2024-12-02T17:28:27,726 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T17:28:27,726 DEBUG [RS:0;3b3375d1a925:40037 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 3b3375d1a925,40037,1733160507453 2024-12-02T17:28:27,726 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,40037,1733160507453' 2024-12-02T17:28:27,726 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T17:28:27,726 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T17:28:27,727 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T17:28:27,727 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T17:28:27,727 DEBUG [RS:0;3b3375d1a925:40037 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 3b3375d1a925,40037,1733160507453 2024-12-02T17:28:27,727 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,40037,1733160507453' 2024-12-02T17:28:27,727 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T17:28:27,727 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T17:28:27,727 DEBUG [RS:0;3b3375d1a925:40037 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T17:28:27,727 INFO [RS:0;3b3375d1a925:40037 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T17:28:27,727 INFO [RS:0;3b3375d1a925:40037 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T17:28:27,769 WARN [3b3375d1a925:39831 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-02T17:28:27,829 INFO [RS:0;3b3375d1a925:40037 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C40037%2C1733160507453, suffix=, logDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453, archiveDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/oldWALs, maxLogs=32 2024-12-02T17:28:27,829 INFO [RS:0;3b3375d1a925:40037 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C40037%2C1733160507453.1733160507829 2024-12-02T17:28:27,834 INFO [RS:0;3b3375d1a925:40037 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.1733160507829 2024-12-02T17:28:27,835 DEBUG [RS:0;3b3375d1a925:40037 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35929:35929),(127.0.0.1/127.0.0.1:40389:40389)] 2024-12-02T17:28:27,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:28:27,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:28:27,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T17:28:28,020 DEBUG [3b3375d1a925:39831 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T17:28:28,020 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:28,021 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,40037,1733160507453, state=OPENING 2024-12-02T17:28:28,023 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T17:28:28,024 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:28,024 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:28:28,025 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:28:28,025 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:28:28,025 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:28:28,025 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,40037,1733160507453}] 2024-12-02T17:28:28,178 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T17:28:28,180 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-15-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:60511, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T17:28:28,183 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T17:28:28,183 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:28:28,185 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C40037%2C1733160507453.meta, suffix=.meta, logDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453, archiveDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/oldWALs, maxLogs=32 2024-12-02T17:28:28,185 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C40037%2C1733160507453.meta.1733160508185.meta 2024-12-02T17:28:28,191 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.meta.1733160508185.meta 2024-12-02T17:28:28,193 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35929:35929),(127.0.0.1/127.0.0.1:40389:40389)] 2024-12-02T17:28:28,197 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:28:28,197 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T17:28:28,197 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T17:28:28,198 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T17:28:28,198 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T17:28:28,198 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:28:28,198 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T17:28:28,198 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T17:28:28,200 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:28:28,201 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:28:28,201 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:28,201 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:28,201 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:28:28,202 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:28:28,202 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:28,202 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:28,203 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:28:28,203 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:28:28,203 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:28,203 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:28,204 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:28:28,204 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:28:28,204 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:28,204 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:28:28,205 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:28:28,205 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740 2024-12-02T17:28:28,206 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740 2024-12-02T17:28:28,207 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:28:28,207 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:28:28,208 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:28:28,209 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:28:28,210 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=809401, jitterRate=0.029207512736320496}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:28:28,210 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T17:28:28,211 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733160508198Writing region info on filesystem at 1733160508198Initializing all the Stores at 1733160508199 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160508199Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160508200 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160508200Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160508200Cleaning up temporary data from old regions at 1733160508207 (+7 ms)Running coprocessor post-open hooks at 1733160508210 (+3 ms)Region opened successfully at 1733160508211 (+1 ms) 2024-12-02T17:28:28,212 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733160508177 2024-12-02T17:28:28,214 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T17:28:28,214 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T17:28:28,215 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:28,216 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,40037,1733160507453, state=OPEN 2024-12-02T17:28:28,220 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:28:28,220 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:28:28,220 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:28:28,220 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:28,220 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:28:28,223 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T17:28:28,223 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,40037,1733160507453 in 195 msec 2024-12-02T17:28:28,225 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T17:28:28,225 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 606 msec 2024-12-02T17:28:28,226 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:28:28,226 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T17:28:28,227 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:28:28,227 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,40037,1733160507453, seqNum=-1] 2024-12-02T17:28:28,227 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:28:28,229 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-15-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:40533, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:28:28,233 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 650 msec 2024-12-02T17:28:28,233 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733160508233, completionTime=-1 2024-12-02T17:28:28,233 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T17:28:28,233 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T17:28:28,235 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=1 2024-12-02T17:28:28,235 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733160568235 2024-12-02T17:28:28,235 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733160628235 2024-12-02T17:28:28,235 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 1 msec 2024-12-02T17:28:28,235 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,39831,1733160507412-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:28,235 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,39831,1733160507412-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:28,235 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,39831,1733160507412-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:28,236 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-3b3375d1a925:39831, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:28,236 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:28,236 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:28,237 DEBUG [master/3b3375d1a925:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T17:28:28,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.757sec 2024-12-02T17:28:28,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T17:28:28,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T17:28:28,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T17:28:28,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T17:28:28,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T17:28:28,239 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,39831,1733160507412-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:28:28,240 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,39831,1733160507412-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T17:28:28,242 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T17:28:28,242 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T17:28:28,242 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,39831,1733160507412-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:28:28,267 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5ce0a24, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:28:28,267 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 3b3375d1a925,39831,-1 for getting cluster id 2024-12-02T17:28:28,267 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T17:28:28,268 DEBUG [HMaster-EventLoopGroup-14-3 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'f03793dc-8a4a-47ee-9dd7-4c3d64891b97' 2024-12-02T17:28:28,268 DEBUG [RPCClient-NioEventLoopGroup-4-4 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T17:28:28,268 DEBUG [RPCClient-NioEventLoopGroup-4-4 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "f03793dc-8a4a-47ee-9dd7-4c3d64891b97" 2024-12-02T17:28:28,269 DEBUG [RPCClient-NioEventLoopGroup-4-4 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@42f2c6cd, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:28:28,269 DEBUG [RPCClient-NioEventLoopGroup-4-4 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [3b3375d1a925,39831,-1] 2024-12-02T17:28:28,269 DEBUG [RPCClient-NioEventLoopGroup-4-4 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T17:28:28,269 DEBUG [RPCClient-NioEventLoopGroup-4-4 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:28:28,270 INFO [HMaster-EventLoopGroup-14-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:39016, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T17:28:28,271 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@470e396e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:28:28,271 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:28:28,272 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,40037,1733160507453, seqNum=-1] 2024-12-02T17:28:28,272 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:28:28,273 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-15-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51472, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:28:28,274 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=3b3375d1a925,39831,1733160507412 2024-12-02T17:28:28,275 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:28:28,277 INFO [Time-limited test {}] master.MasterRpcServices(567): Client=null/null set balanceSwitch=false 2024-12-02T17:28:28,277 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-02T17:28:28,278 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] client.AsyncConnectionImpl(321): The fetched master address is 3b3375d1a925,39831,1733160507412 2024-12-02T17:28:28,278 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@3d8d3080 2024-12-02T17:28:28,278 DEBUG [RPCClient-NioEventLoopGroup-4-5 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T17:28:28,279 INFO [HMaster-EventLoopGroup-14-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:39026, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T17:28:28,279 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39831 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T17:28:28,279 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39831 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T17:28:28,279 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39831 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestLogRolling-testLogRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:28:28,280 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39831 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestLogRolling-testLogRolling 2024-12-02T17:28:28,282 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T17:28:28,282 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:28,282 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39831 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRolling" procId is: 4 2024-12-02T17:28:28,283 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39831 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:28:28,283 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T17:28:28,289 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741835_1011 (size=381) 2024-12-02T17:28:28,289 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741835_1011 (size=381) 2024-12-02T17:28:28,291 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 2bd92dcfc9073d608b4ecf1e0d6b55d5, NAME => 'TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec 2024-12-02T17:28:28,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741836_1012 (size=64) 2024-12-02T17:28:28,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741836_1012 (size=64) 2024-12-02T17:28:28,297 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(898): Instantiated TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:28:28,297 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1722): Closing 2bd92dcfc9073d608b4ecf1e0d6b55d5, disabling compactions & flushes 2024-12-02T17:28:28,297 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1755): Closing region TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:28,297 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:28,297 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. after waiting 0 ms 2024-12-02T17:28:28,297 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:28,297 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1973): Closed TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:28,297 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1676): Region close journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: Waiting for close lock at 1733160508297Disabling compacts and flushes for region at 1733160508297Disabling writes for close at 1733160508297Writing region close event to WAL at 1733160508297Closed at 1733160508297 2024-12-02T17:28:28,298 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T17:28:28,298 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.","families":{"info":[{"qualifier":"regioninfo","vlen":63,"tag":[],"timestamp":"1733160508298"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733160508298"}]},"ts":"1733160508298"} 2024-12-02T17:28:28,301 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-02T17:28:28,302 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T17:28:28,302 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160508302"}]},"ts":"1733160508302"} 2024-12-02T17:28:28,304 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testLogRolling, state=ENABLING in hbase:meta 2024-12-02T17:28:28,304 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, ASSIGN}] 2024-12-02T17:28:28,305 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, ASSIGN 2024-12-02T17:28:28,306 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, ASSIGN; state=OFFLINE, location=3b3375d1a925,40037,1733160507453; forceNewPlan=false, retain=false 2024-12-02T17:28:28,457 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=2bd92dcfc9073d608b4ecf1e0d6b55d5, regionState=OPENING, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:28,459 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-15-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, ASSIGN because future has completed 2024-12-02T17:28:28,459 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453}] 2024-12-02T17:28:28,616 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:28,616 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 2bd92dcfc9073d608b4ecf1e0d6b55d5, NAME => 'TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:28:28,617 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,617 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:28:28,617 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,617 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,618 INFO [StoreOpener-2bd92dcfc9073d608b4ecf1e0d6b55d5-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,619 INFO [StoreOpener-2bd92dcfc9073d608b4ecf1e0d6b55d5-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 2bd92dcfc9073d608b4ecf1e0d6b55d5 columnFamilyName info 2024-12-02T17:28:28,619 DEBUG [StoreOpener-2bd92dcfc9073d608b4ecf1e0d6b55d5-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:28,619 INFO [StoreOpener-2bd92dcfc9073d608b4ecf1e0d6b55d5-1 {}] regionserver.HStore(327): Store=2bd92dcfc9073d608b4ecf1e0d6b55d5/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:28:28,620 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,620 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,620 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,621 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,621 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,622 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,624 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:28:28,624 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 2bd92dcfc9073d608b4ecf1e0d6b55d5; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=800924, jitterRate=0.018427729606628418}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T17:28:28,624 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:28,625 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: Running coprocessor pre-open hook at 1733160508617Writing region info on filesystem at 1733160508617Initializing all the Stores at 1733160508617Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160508617Cleaning up temporary data from old regions at 1733160508621 (+4 ms)Running coprocessor post-open hooks at 1733160508624 (+3 ms)Region opened successfully at 1733160508625 (+1 ms) 2024-12-02T17:28:28,626 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., pid=6, masterSystemTime=1733160508612 2024-12-02T17:28:28,628 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:28,628 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:28,629 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=2bd92dcfc9073d608b4ecf1e0d6b55d5, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:28,631 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-15-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 because future has completed 2024-12-02T17:28:28,635 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T17:28:28,635 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 in 173 msec 2024-12-02T17:28:28,637 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T17:28:28,637 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, ASSIGN in 331 msec 2024-12-02T17:28:28,638 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T17:28:28,639 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestLogRolling-testLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733160508638"}]},"ts":"1733160508638"} 2024-12-02T17:28:28,641 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestLogRolling-testLogRolling, state=ENABLED in hbase:meta 2024-12-02T17:28:28,642 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T17:28:28,643 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestLogRolling-testLogRolling in 363 msec 2024-12-02T17:28:31,271 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,271 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,272 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,272 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,272 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,272 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,272 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,273 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,285 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,285 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,285 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,285 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,285 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,286 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,288 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,288 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,289 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,290 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,794 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:28:31,795 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,795 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,795 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,795 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,795 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,795 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,796 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,796 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,811 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,811 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,811 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,812 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,812 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,812 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,815 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,815 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,815 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:31,818 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:33,693 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T17:28:33,694 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRolling' 2024-12-02T17:28:37,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:28:37,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-02T17:28:37,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:28:37,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-02T17:28:37,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRolling 2024-12-02T17:28:37,867 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRolling Metrics about Tables on a single HBase RegionServer 2024-12-02T17:28:38,300 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39831 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-02T17:28:38,300 INFO [RPCClient-NioEventLoopGroup-4-7 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestLogRolling-testLogRolling completed 2024-12-02T17:28:38,300 DEBUG [Time-limited test {}] hbase.ClientMetaTableAccessor(255): Scanning META starting at row=TestLogRolling-testLogRolling,, stopping at row=TestLogRolling-testLogRolling ,, for max=2147483647 with caching=100 2024-12-02T17:28:38,303 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2234): Found 1 regions for table TestLogRolling-testLogRolling 2024-12-02T17:28:38,303 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(2240): firstRegionName=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:38,305 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestLogRolling-testLogRolling', row='row0001', locateType=CURRENT is [region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2] 2024-12-02T17:28:38,316 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:38,317 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:28:38,333 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/78c2a5c8cfec44f4a66a9aeceaaccd6b is 1080, key is row0001/info:/1733160518306/Put/seqid=0 2024-12-02T17:28:38,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741837_1013 (size=12509) 2024-12-02T17:28:38,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741837_1013 (size=12509) 2024-12-02T17:28:38,352 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/78c2a5c8cfec44f4a66a9aeceaaccd6b 2024-12-02T17:28:38,351 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(5310): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:3.0.0-beta-2-SNAPSHOT] 2024-12-02T17:28:38,356 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] ipc.CallRunner(138): callId: 35 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.3:51472 deadline: 1733160528351, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:38,358 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/78c2a5c8cfec44f4a66a9aeceaaccd6b as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/78c2a5c8cfec44f4a66a9aeceaaccd6b 2024-12-02T17:28:38,364 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/78c2a5c8cfec44f4a66a9aeceaaccd6b, entries=7, sequenceid=11, filesize=12.2 K 2024-12-02T17:28:38,365 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for 2bd92dcfc9073d608b4ecf1e0d6b55d5 in 49ms, sequenceid=11, compaction requested=false 2024-12-02T17:28:38,365 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: 2024-12-02T17:28:38,379 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(64): Try updating region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 , the old value is region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2, error=org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:28:38,379 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(72): The actual exception when updating region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 is org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:28:38,380 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(76): Will not update region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 because the exception is null or not the one we care about 2024-12-02T17:28:48,460 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:48,461 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5 1/1 column families, dataSize=24.17 KB heapSize=26.13 KB 2024-12-02T17:28:48,466 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/892eb68a23d0432ab818790535bf2fc8 is 1080, key is row0008/info:/1733160518317/Put/seqid=0 2024-12-02T17:28:48,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741838_1014 (size=29761) 2024-12-02T17:28:48,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741838_1014 (size=29761) 2024-12-02T17:28:48,471 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.17 KB at sequenceid=37 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/892eb68a23d0432ab818790535bf2fc8 2024-12-02T17:28:48,477 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/892eb68a23d0432ab818790535bf2fc8 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8 2024-12-02T17:28:48,483 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8, entries=23, sequenceid=37, filesize=29.1 K 2024-12-02T17:28:48,484 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~24.17 KB/24748, heapSize ~26.11 KB/26736, currentSize=2.10 KB/2152 for 2bd92dcfc9073d608b4ecf1e0d6b55d5 in 24ms, sequenceid=37, compaction requested=false 2024-12-02T17:28:48,484 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: 2024-12-02T17:28:48,484 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=41.3 K, sizeToCheck=16.0 K 2024-12-02T17:28:48,484 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:48,484 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8 because midkey is the same as first or last row 2024-12-02T17:28:50,473 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:50,473 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:28:50,478 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/b2ce90b3a99f4fe1a3b70216ad4457ee is 1080, key is row0031/info:/1733160528462/Put/seqid=0 2024-12-02T17:28:50,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741839_1015 (size=12509) 2024-12-02T17:28:50,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741839_1015 (size=12509) 2024-12-02T17:28:50,489 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=47 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/b2ce90b3a99f4fe1a3b70216ad4457ee 2024-12-02T17:28:50,496 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/b2ce90b3a99f4fe1a3b70216ad4457ee as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/b2ce90b3a99f4fe1a3b70216ad4457ee 2024-12-02T17:28:50,502 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/b2ce90b3a99f4fe1a3b70216ad4457ee, entries=7, sequenceid=47, filesize=12.2 K 2024-12-02T17:28:50,502 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=17.86 KB/18292 for 2bd92dcfc9073d608b4ecf1e0d6b55d5 in 29ms, sequenceid=47, compaction requested=true 2024-12-02T17:28:50,502 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: 2024-12-02T17:28:50,503 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=53.5 K, sizeToCheck=16.0 K 2024-12-02T17:28:50,503 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:50,503 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8 because midkey is the same as first or last row 2024-12-02T17:28:50,503 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 2bd92dcfc9073d608b4ecf1e0d6b55d5:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:28:50,503 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:28:50,503 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:28:50,504 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:50,504 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5 1/1 column families, dataSize=18.91 KB heapSize=20.50 KB 2024-12-02T17:28:50,504 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 54779 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:28:50,504 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): 2bd92dcfc9073d608b4ecf1e0d6b55d5/info is initiating minor compaction (all files) 2024-12-02T17:28:50,504 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of 2bd92dcfc9073d608b4ecf1e0d6b55d5/info in TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:50,504 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/78c2a5c8cfec44f4a66a9aeceaaccd6b, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/b2ce90b3a99f4fe1a3b70216ad4457ee] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp, totalSize=53.5 K 2024-12-02T17:28:50,505 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 78c2a5c8cfec44f4a66a9aeceaaccd6b, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=11, earliestPutTs=1733160518306 2024-12-02T17:28:50,505 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 892eb68a23d0432ab818790535bf2fc8, keycount=23, bloomtype=ROW, size=29.1 K, encoding=NONE, compression=NONE, seqNum=37, earliestPutTs=1733160518317 2024-12-02T17:28:50,506 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting b2ce90b3a99f4fe1a3b70216ad4457ee, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=47, earliestPutTs=1733160528462 2024-12-02T17:28:50,508 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/57b2b75eb1b846b480a0d23889813b11 is 1080, key is row0038/info:/1733160530474/Put/seqid=0 2024-12-02T17:28:50,524 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 2bd92dcfc9073d608b4ecf1e0d6b55d5#info#compaction#59 average throughput is 18.98 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:28:50,524 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/3f5299b7a495432383bb15ba6c156aba is 1080, key is row0001/info:/1733160518306/Put/seqid=0 2024-12-02T17:28:50,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741840_1016 (size=24376) 2024-12-02T17:28:50,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741840_1016 (size=24376) 2024-12-02T17:28:50,526 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=18.91 KB at sequenceid=68 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/57b2b75eb1b846b480a0d23889813b11 2024-12-02T17:28:50,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741841_1017 (size=44978) 2024-12-02T17:28:50,531 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741841_1017 (size=44978) 2024-12-02T17:28:50,531 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/57b2b75eb1b846b480a0d23889813b11 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/57b2b75eb1b846b480a0d23889813b11 2024-12-02T17:28:50,536 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/3f5299b7a495432383bb15ba6c156aba as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba 2024-12-02T17:28:50,537 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/57b2b75eb1b846b480a0d23889813b11, entries=18, sequenceid=68, filesize=23.8 K 2024-12-02T17:28:50,538 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~18.91 KB/19368, heapSize ~20.48 KB/20976, currentSize=9.46 KB/9684 for 2bd92dcfc9073d608b4ecf1e0d6b55d5 in 33ms, sequenceid=68, compaction requested=false 2024-12-02T17:28:50,538 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: 2024-12-02T17:28:50,538 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=77.3 K, sizeToCheck=16.0 K 2024-12-02T17:28:50,538 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:50,538 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8 because midkey is the same as first or last row 2024-12-02T17:28:50,542 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in 2bd92dcfc9073d608b4ecf1e0d6b55d5/info of 2bd92dcfc9073d608b4ecf1e0d6b55d5 into 3f5299b7a495432383bb15ba6c156aba(size=43.9 K), total size for store is 67.7 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: 2024-12-02T17:28:50,542 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., storeName=2bd92dcfc9073d608b4ecf1e0d6b55d5/info, priority=13, startTime=1733160530503; duration=0sec 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=67.7 K, sizeToCheck=16.0 K 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba because midkey is the same as first or last row 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=67.7 K, sizeToCheck=16.0 K 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba because midkey is the same as first or last row 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=67.7 K, sizeToCheck=16.0 K 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba because midkey is the same as first or last row 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:28:50,542 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 2bd92dcfc9073d608b4ecf1e0d6b55d5:info 2024-12-02T17:28:52,522 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,522 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5 1/1 column families, dataSize=10.51 KB heapSize=11.50 KB 2024-12-02T17:28:52,526 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/302acc96c56246beab4ee9b3d690a7a8 is 1080, key is row0056/info:/1733160530505/Put/seqid=0 2024-12-02T17:28:52,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741842_1018 (size=15740) 2024-12-02T17:28:52,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741842_1018 (size=15740) 2024-12-02T17:28:52,535 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=10.51 KB at sequenceid=82 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/302acc96c56246beab4ee9b3d690a7a8 2024-12-02T17:28:52,541 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/302acc96c56246beab4ee9b3d690a7a8 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/302acc96c56246beab4ee9b3d690a7a8 2024-12-02T17:28:52,546 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/302acc96c56246beab4ee9b3d690a7a8, entries=10, sequenceid=82, filesize=15.4 K 2024-12-02T17:28:52,547 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~10.51 KB/10760, heapSize ~11.48 KB/11760, currentSize=14.71 KB/15064 for 2bd92dcfc9073d608b4ecf1e0d6b55d5 in 25ms, sequenceid=82, compaction requested=true 2024-12-02T17:28:52,547 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: 2024-12-02T17:28:52,547 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=83.1 K, sizeToCheck=16.0 K 2024-12-02T17:28:52,547 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:52,548 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba because midkey is the same as first or last row 2024-12-02T17:28:52,548 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 2bd92dcfc9073d608b4ecf1e0d6b55d5:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:28:52,548 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:28:52,548 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:28:52,549 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 85094 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:28:52,549 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): 2bd92dcfc9073d608b4ecf1e0d6b55d5/info is initiating minor compaction (all files) 2024-12-02T17:28:52,549 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of 2bd92dcfc9073d608b4ecf1e0d6b55d5/info in TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:52,549 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/57b2b75eb1b846b480a0d23889813b11, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/302acc96c56246beab4ee9b3d690a7a8] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp, totalSize=83.1 K 2024-12-02T17:28:52,549 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,549 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5 1/1 column families, dataSize=16.81 KB heapSize=18.25 KB 2024-12-02T17:28:52,550 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 3f5299b7a495432383bb15ba6c156aba, keycount=37, bloomtype=ROW, size=43.9 K, encoding=NONE, compression=NONE, seqNum=47, earliestPutTs=1733160518306 2024-12-02T17:28:52,550 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 57b2b75eb1b846b480a0d23889813b11, keycount=18, bloomtype=ROW, size=23.8 K, encoding=NONE, compression=NONE, seqNum=68, earliestPutTs=1733160530474 2024-12-02T17:28:52,550 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 302acc96c56246beab4ee9b3d690a7a8, keycount=10, bloomtype=ROW, size=15.4 K, encoding=NONE, compression=NONE, seqNum=82, earliestPutTs=1733160530505 2024-12-02T17:28:52,553 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/e1c61ca3dbfd4c3fade405d03f737936 is 1080, key is row0066/info:/1733160532523/Put/seqid=0 2024-12-02T17:28:52,570 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741843_1019 (size=22222) 2024-12-02T17:28:52,571 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(5310): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:3.0.0-beta-2-SNAPSHOT] 2024-12-02T17:28:52,571 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741843_1019 (size=22222) 2024-12-02T17:28:52,571 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] ipc.CallRunner(138): callId: 101 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.3:51472 deadline: 1733160542571, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:52,572 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(64): Try updating region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 , the old value is region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2, error=org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:28:52,572 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(72): The actual exception when updating region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 is org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:28:52,572 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(76): Will not update region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 because the exception is null or not the one we care about 2024-12-02T17:28:52,572 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=16.81 KB at sequenceid=101 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/e1c61ca3dbfd4c3fade405d03f737936 2024-12-02T17:28:52,575 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 2bd92dcfc9073d608b4ecf1e0d6b55d5#info#compaction#62 average throughput is 22.23 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:28:52,576 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/c66a83493e5a4dfcada1bd023a2ae935 is 1080, key is row0001/info:/1733160518306/Put/seqid=0 2024-12-02T17:28:52,577 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/e1c61ca3dbfd4c3fade405d03f737936 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/e1c61ca3dbfd4c3fade405d03f737936 2024-12-02T17:28:52,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741844_1020 (size=75378) 2024-12-02T17:28:52,580 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741844_1020 (size=75378) 2024-12-02T17:28:52,582 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/e1c61ca3dbfd4c3fade405d03f737936, entries=16, sequenceid=101, filesize=21.7 K 2024-12-02T17:28:52,583 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~16.81 KB/17216, heapSize ~18.23 KB/18672, currentSize=13.66 KB/13988 for 2bd92dcfc9073d608b4ecf1e0d6b55d5 in 34ms, sequenceid=101, compaction requested=false 2024-12-02T17:28:52,583 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: 2024-12-02T17:28:52,583 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=104.8 K, sizeToCheck=16.0 K 2024-12-02T17:28:52,583 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:52,583 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba because midkey is the same as first or last row 2024-12-02T17:28:52,585 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/c66a83493e5a4dfcada1bd023a2ae935 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935 2024-12-02T17:28:52,590 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in 2bd92dcfc9073d608b4ecf1e0d6b55d5/info of 2bd92dcfc9073d608b4ecf1e0d6b55d5 into c66a83493e5a4dfcada1bd023a2ae935(size=73.6 K), total size for store is 95.3 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:28:52,590 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: 2024-12-02T17:28:52,590 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., storeName=2bd92dcfc9073d608b4ecf1e0d6b55d5/info, priority=13, startTime=1733160532548; duration=0sec 2024-12-02T17:28:52,590 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=95.3 K, sizeToCheck=16.0 K 2024-12-02T17:28:52,590 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:52,590 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=95.3 K, sizeToCheck=16.0 K 2024-12-02T17:28:52,590 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:52,590 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(101): Should split because region size is big enough sumSize=95.3 K, sizeToCheck=16.0 K 2024-12-02T17:28:52,590 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T17:28:52,591 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit(239): Splitting TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:28:52,591 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:28:52,591 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 2bd92dcfc9073d608b4ecf1e0d6b55d5:info 2024-12-02T17:28:52,592 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39831 {}] assignment.AssignmentManager(1363): Split request from 3b3375d1a925,40037,1733160507453, parent={ENCODED => 2bd92dcfc9073d608b4ecf1e0d6b55d5, NAME => 'TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.', STARTKEY => '', ENDKEY => ''}, splitKey=row0062 2024-12-02T17:28:52,597 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39831 {}] assignment.SplitTableRegionProcedure(223): Splittable=true state=OPEN, location=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:52,601 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39831 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE, hasLock=false; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=2bd92dcfc9073d608b4ecf1e0d6b55d5, daughterA=cd0c090bdce692f9dab734ec9b0dfe67, daughterB=b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:52,602 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=7, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE, hasLock=false; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=2bd92dcfc9073d608b4ecf1e0d6b55d5, daughterA=cd0c090bdce692f9dab734ec9b0dfe67, daughterB=b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:52,602 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=7, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE, hasLock=false; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=2bd92dcfc9073d608b4ecf1e0d6b55d5, daughterA=cd0c090bdce692f9dab734ec9b0dfe67, daughterB=b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:52,602 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=7, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE, hasLock=false; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=2bd92dcfc9073d608b4ecf1e0d6b55d5, daughterA=cd0c090bdce692f9dab734ec9b0dfe67, daughterB=b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:52,608 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, UNASSIGN}] 2024-12-02T17:28:52,609 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=8, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, UNASSIGN 2024-12-02T17:28:52,611 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=8 updating hbase:meta row=2bd92dcfc9073d608b4ecf1e0d6b55d5, regionState=CLOSING, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:52,612 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-15-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=8, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, UNASSIGN because future has completed 2024-12-02T17:28:52,613 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(375): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2024-12-02T17:28:52,613 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=9, ppid=8, state=RUNNABLE, hasLock=false; CloseRegionProcedure 2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453}] 2024-12-02T17:28:52,769 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] handler.UnassignRegionHandler(122): Close 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,769 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] handler.UnassignRegionHandler(136): Unassign region: split region: true: evictCache: true 2024-12-02T17:28:52,769 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(1722): Closing 2bd92dcfc9073d608b4ecf1e0d6b55d5, disabling compactions & flushes 2024-12-02T17:28:52,769 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(1755): Closing region TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:52,769 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:52,769 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. after waiting 0 ms 2024-12-02T17:28:52,769 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:52,770 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(2902): Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5 1/1 column families, dataSize=13.66 KB heapSize=14.88 KB 2024-12-02T17:28:52,773 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/c5f21f4850c9479aba2e5dc3dec7b9a2 is 1080, key is row0082/info:/1733160532550/Put/seqid=0 2024-12-02T17:28:52,777 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741845_1021 (size=18987) 2024-12-02T17:28:52,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741845_1021 (size=18987) 2024-12-02T17:28:52,782 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=13.66 KB at sequenceid=118 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/c5f21f4850c9479aba2e5dc3dec7b9a2 2024-12-02T17:28:52,788 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/.tmp/info/c5f21f4850c9479aba2e5dc3dec7b9a2 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c5f21f4850c9479aba2e5dc3dec7b9a2 2024-12-02T17:28:52,793 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c5f21f4850c9479aba2e5dc3dec7b9a2, entries=13, sequenceid=118, filesize=18.5 K 2024-12-02T17:28:52,794 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(3140): Finished flush of dataSize ~13.66 KB/13988, heapSize ~14.86 KB/15216, currentSize=0 B/0 for 2bd92dcfc9073d608b4ecf1e0d6b55d5 in 25ms, sequenceid=118, compaction requested=true 2024-12-02T17:28:52,795 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.-1 {}] regionserver.HStore(2317): Moving the files [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/78c2a5c8cfec44f4a66a9aeceaaccd6b, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/b2ce90b3a99f4fe1a3b70216ad4457ee, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/57b2b75eb1b846b480a0d23889813b11, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/302acc96c56246beab4ee9b3d690a7a8] to archive 2024-12-02T17:28:52,796 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T17:28:52,798 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/78c2a5c8cfec44f4a66a9aeceaaccd6b to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/78c2a5c8cfec44f4a66a9aeceaaccd6b 2024-12-02T17:28:52,799 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/892eb68a23d0432ab818790535bf2fc8 2024-12-02T17:28:52,800 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/3f5299b7a495432383bb15ba6c156aba 2024-12-02T17:28:52,801 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/b2ce90b3a99f4fe1a3b70216ad4457ee to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/b2ce90b3a99f4fe1a3b70216ad4457ee 2024-12-02T17:28:52,802 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/57b2b75eb1b846b480a0d23889813b11 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/57b2b75eb1b846b480a0d23889813b11 2024-12-02T17:28:52,803 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/302acc96c56246beab4ee9b3d690a7a8 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/302acc96c56246beab4ee9b3d690a7a8 2024-12-02T17:28:52,809 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/recovered.edits/121.seqid, newMaxSeqId=121, maxSeqId=1 2024-12-02T17:28:52,810 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(1973): Closed TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. 2024-12-02T17:28:52,810 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] regionserver.HRegion(1676): Region close journal for 2bd92dcfc9073d608b4ecf1e0d6b55d5: Waiting for close lock at 1733160532769Running coprocessor pre-close hooks at 1733160532769Disabling compacts and flushes for region at 1733160532769Disabling writes for close at 1733160532769Obtaining lock to block concurrent updates at 1733160532770 (+1 ms)Preparing flush snapshotting stores in 2bd92dcfc9073d608b4ecf1e0d6b55d5 at 1733160532770Finished memstore snapshotting TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., syncing WAL and waiting on mvcc, flushsize=dataSize=13988, getHeapSize=15216, getOffHeapSize=0, getCellsCount=13 at 1733160532770Flushing stores of TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. at 1733160532770Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5/info: creating writer at 1733160532770Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5/info: appending metadata at 1733160532773 (+3 ms)Flushing 2bd92dcfc9073d608b4ecf1e0d6b55d5/info: closing flushed file at 1733160532773Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@70fa208e: reopening flushed file at 1733160532787 (+14 ms)Finished flush of dataSize ~13.66 KB/13988, heapSize ~14.86 KB/15216, currentSize=0 B/0 for 2bd92dcfc9073d608b4ecf1e0d6b55d5 in 25ms, sequenceid=118, compaction requested=true at 1733160532794 (+7 ms)Writing region close event to WAL at 1733160532806 (+12 ms)Running coprocessor post-close hooks at 1733160532809 (+3 ms)Closed at 1733160532809 2024-12-02T17:28:52,812 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION, pid=9}] handler.UnassignRegionHandler(157): Closed 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,813 INFO [PEWorker-1 {}] assignment.RegionStateStore(223): pid=8 updating hbase:meta row=2bd92dcfc9073d608b4ecf1e0d6b55d5, regionState=CLOSED 2024-12-02T17:28:52,814 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-15-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=9, ppid=8, state=RUNNABLE, hasLock=false; CloseRegionProcedure 2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 because future has completed 2024-12-02T17:28:52,817 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=9, resume processing ppid=8 2024-12-02T17:28:52,817 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=9, ppid=8, state=SUCCESS, hasLock=false; CloseRegionProcedure 2bd92dcfc9073d608b4ecf1e0d6b55d5, server=3b3375d1a925,40037,1733160507453 in 202 msec 2024-12-02T17:28:52,819 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-02T17:28:52,819 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, UNASSIGN in 209 msec 2024-12-02T17:28:52,827 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:52,831 INFO [PEWorker-4 {}] assignment.SplitTableRegionProcedure(728): pid=7 splitting 3 storefiles, region=2bd92dcfc9073d608b4ecf1e0d6b55d5, threads=3 2024-12-02T17:28:52,833 DEBUG [StoreFileSplitter-pool-2 {}] assignment.SplitTableRegionProcedure(823): pid=7 splitting started for store file: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/e1c61ca3dbfd4c3fade405d03f737936 for region: 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,833 DEBUG [StoreFileSplitter-pool-0 {}] assignment.SplitTableRegionProcedure(823): pid=7 splitting started for store file: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c5f21f4850c9479aba2e5dc3dec7b9a2 for region: 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,833 DEBUG [StoreFileSplitter-pool-1 {}] assignment.SplitTableRegionProcedure(823): pid=7 splitting started for store file: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935 for region: 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,843 DEBUG [StoreFileSplitter-pool-2 {}] regionserver.HRegionFileSystem(650): Will create HFileLink file for hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/e1c61ca3dbfd4c3fade405d03f737936, top=true 2024-12-02T17:28:52,847 DEBUG [StoreFileSplitter-pool-0 {}] regionserver.HRegionFileSystem(650): Will create HFileLink file for hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c5f21f4850c9479aba2e5dc3dec7b9a2, top=true 2024-12-02T17:28:52,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741846_1022 (size=27) 2024-12-02T17:28:52,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741846_1022 (size=27) 2024-12-02T17:28:52,851 INFO [StoreFileSplitter-pool-2 {}] regionserver.HRegionFileSystem(691): Created linkFile:hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-e1c61ca3dbfd4c3fade405d03f737936 for child: b10b696b6a97e7d9d531081df90f6c2a, parent: 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,852 INFO [StoreFileSplitter-pool-0 {}] regionserver.HRegionFileSystem(691): Created linkFile:hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-c5f21f4850c9479aba2e5dc3dec7b9a2 for child: b10b696b6a97e7d9d531081df90f6c2a, parent: 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,852 DEBUG [StoreFileSplitter-pool-2 {}] assignment.SplitTableRegionProcedure(834): pid=7 splitting complete for store file: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/e1c61ca3dbfd4c3fade405d03f737936 for region: 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,852 DEBUG [StoreFileSplitter-pool-0 {}] assignment.SplitTableRegionProcedure(834): pid=7 splitting complete for store file: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c5f21f4850c9479aba2e5dc3dec7b9a2 for region: 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,862 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741847_1023 (size=27) 2024-12-02T17:28:52,862 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741847_1023 (size=27) 2024-12-02T17:28:52,862 DEBUG [StoreFileSplitter-pool-1 {}] assignment.SplitTableRegionProcedure(834): pid=7 splitting complete for store file: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935 for region: 2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:28:52,864 DEBUG [PEWorker-4 {}] assignment.SplitTableRegionProcedure(802): pid=7 split storefiles for region 2bd92dcfc9073d608b4ecf1e0d6b55d5 Daughter A: [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5] storefiles, Daughter B: [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-c5f21f4850c9479aba2e5dc3dec7b9a2, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-e1c61ca3dbfd4c3fade405d03f737936] storefiles. 2024-12-02T17:28:52,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741848_1024 (size=71) 2024-12-02T17:28:52,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741848_1024 (size=71) 2024-12-02T17:28:52,873 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:52,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741849_1025 (size=71) 2024-12-02T17:28:52,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741849_1025 (size=71) 2024-12-02T17:28:52,886 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:52,896 DEBUG [PEWorker-4 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/recovered.edits/121.seqid, newMaxSeqId=121, maxSeqId=-1 2024-12-02T17:28:52,898 DEBUG [PEWorker-4 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/recovered.edits/121.seqid, newMaxSeqId=121, maxSeqId=-1 2024-12-02T17:28:52,901 DEBUG [PEWorker-4 {}] assignment.RegionStateStore(723): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5.","families":{"info":[{"qualifier":"regioninfo","vlen":63,"tag":[],"timestamp":"1733160532900"},{"qualifier":"splitA","vlen":70,"tag":[],"timestamp":"1733160532900"},{"qualifier":"splitB","vlen":70,"tag":[],"timestamp":"1733160532900"}]},"ts":"1733160532900"} 2024-12-02T17:28:52,901 DEBUG [PEWorker-4 {}] assignment.RegionStateStore(723): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67.","families":{"info":[{"qualifier":"regioninfo","vlen":70,"tag":[],"timestamp":"1733160532900"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733160532900"},{"qualifier":"seqnumDuringOpen","vlen":8,"tag":[],"timestamp":"1733160532900"}]},"ts":"1733160532900"} 2024-12-02T17:28:52,901 DEBUG [PEWorker-4 {}] assignment.RegionStateStore(723): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.","families":{"info":[{"qualifier":"regioninfo","vlen":70,"tag":[],"timestamp":"1733160532900"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733160532900"},{"qualifier":"seqnumDuringOpen","vlen":8,"tag":[],"timestamp":"1733160532900"}]},"ts":"1733160532900"} 2024-12-02T17:28:52,918 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=10, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=cd0c090bdce692f9dab734ec9b0dfe67, ASSIGN}, {pid=11, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=b10b696b6a97e7d9d531081df90f6c2a, ASSIGN}] 2024-12-02T17:28:52,920 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=10, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=cd0c090bdce692f9dab734ec9b0dfe67, ASSIGN 2024-12-02T17:28:52,920 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=11, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=b10b696b6a97e7d9d531081df90f6c2a, ASSIGN 2024-12-02T17:28:52,920 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=10, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=cd0c090bdce692f9dab734ec9b0dfe67, ASSIGN; state=SPLITTING_NEW, location=3b3375d1a925,40037,1733160507453; forceNewPlan=false, retain=false 2024-12-02T17:28:52,921 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(269): Starting pid=11, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=b10b696b6a97e7d9d531081df90f6c2a, ASSIGN; state=SPLITTING_NEW, location=3b3375d1a925,40037,1733160507453; forceNewPlan=false, retain=false 2024-12-02T17:28:53,071 INFO [PEWorker-2 {}] assignment.RegionStateStore(223): pid=11 updating hbase:meta row=b10b696b6a97e7d9d531081df90f6c2a, regionState=OPENING, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:53,071 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=10 updating hbase:meta row=cd0c090bdce692f9dab734ec9b0dfe67, regionState=OPENING, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:53,073 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-15-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=11, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=b10b696b6a97e7d9d531081df90f6c2a, ASSIGN because future has completed 2024-12-02T17:28:53,073 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=12, ppid=11, state=RUNNABLE, hasLock=false; OpenRegionProcedure b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453}] 2024-12-02T17:28:53,074 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-15-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=10, ppid=7, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=cd0c090bdce692f9dab734ec9b0dfe67, ASSIGN because future has completed 2024-12-02T17:28:53,074 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=13, ppid=10, state=RUNNABLE, hasLock=false; OpenRegionProcedure cd0c090bdce692f9dab734ec9b0dfe67, server=3b3375d1a925,40037,1733160507453}] 2024-12-02T17:28:53,228 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] handler.AssignRegionHandler(132): Open TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:28:53,229 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(7752): Opening region: {ENCODED => cd0c090bdce692f9dab734ec9b0dfe67, NAME => 'TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67.', STARTKEY => '', ENDKEY => 'row0062'} 2024-12-02T17:28:53,229 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,229 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(898): Instantiated TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:28:53,229 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(7794): checking encryption for cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,229 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(7797): checking classloading for cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,230 INFO [StoreOpener-cd0c090bdce692f9dab734ec9b0dfe67-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,231 INFO [StoreOpener-cd0c090bdce692f9dab734ec9b0dfe67-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region cd0c090bdce692f9dab734ec9b0dfe67 columnFamilyName info 2024-12-02T17:28:53,231 DEBUG [StoreOpener-cd0c090bdce692f9dab734ec9b0dfe67-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:53,241 DEBUG [StoreOpener-cd0c090bdce692f9dab734ec9b0dfe67-1 {}] regionserver.StoreEngine(278): loaded hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5->hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935-bottom 2024-12-02T17:28:53,242 INFO [StoreOpener-cd0c090bdce692f9dab734ec9b0dfe67-1 {}] regionserver.HStore(327): Store=cd0c090bdce692f9dab734ec9b0dfe67/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:28:53,242 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(1038): replaying wal for cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,243 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,244 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,244 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(1048): stopping wal replay for cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,244 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(1060): Cleaning up temporary data for cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,246 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(1093): writing seq id for cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,246 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(1114): Opened cd0c090bdce692f9dab734ec9b0dfe67; next sequenceid=122; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=755344, jitterRate=-0.0395309180021286}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T17:28:53,246 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(1122): Running coprocessor post-open hooks for cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:28:53,247 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegion(1006): Region open journal for cd0c090bdce692f9dab734ec9b0dfe67: Running coprocessor pre-open hook at 1733160533229Writing region info on filesystem at 1733160533229Initializing all the Stores at 1733160533230 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160533230Cleaning up temporary data from old regions at 1733160533244 (+14 ms)Running coprocessor post-open hooks at 1733160533246 (+2 ms)Region opened successfully at 1733160533247 (+1 ms) 2024-12-02T17:28:53,248 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegionServer(2236): Post open deploy tasks for TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67., pid=13, masterSystemTime=1733160533225 2024-12-02T17:28:53,248 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.CompactSplit(403): Add compact mark for store cd0c090bdce692f9dab734ec9b0dfe67:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:28:53,248 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 1 store files, 0 compacting, 1 eligible, 16 blocking 2024-12-02T17:28:53,248 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: Opening Region; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:28:53,249 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1527): Keeping/Overriding Compaction request priority to -2147482648 for CF info since it belongs to recently split daughter region TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:28:53,249 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): cd0c090bdce692f9dab734ec9b0dfe67/info is initiating minor compaction (all files) 2024-12-02T17:28:53,249 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of cd0c090bdce692f9dab734ec9b0dfe67/info in TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:28:53,249 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5->hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935-bottom] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/.tmp, totalSize=73.6 K 2024-12-02T17:28:53,249 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5, keycount=32, bloomtype=ROW, size=73.6 K, encoding=NONE, compression=NONE, seqNum=82, earliestPutTs=1733160518306 2024-12-02T17:28:53,250 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] regionserver.HRegionServer(2266): Finished post open deploy task for TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:28:53,250 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=13}] handler.AssignRegionHandler(153): Opened TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:28:53,250 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] handler.AssignRegionHandler(132): Open TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:28:53,250 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(7752): Opening region: {ENCODED => b10b696b6a97e7d9d531081df90f6c2a, NAME => 'TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.', STARTKEY => 'row0062', ENDKEY => ''} 2024-12-02T17:28:53,250 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,251 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(898): Instantiated TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:28:53,251 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(7794): checking encryption for b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,251 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(7797): checking classloading for b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,251 INFO [PEWorker-2 {}] assignment.RegionStateStore(223): pid=10 updating hbase:meta row=cd0c090bdce692f9dab734ec9b0dfe67, regionState=OPEN, openSeqNum=122, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:53,252 INFO [StoreOpener-b10b696b6a97e7d9d531081df90f6c2a-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,253 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on 1588230740 2024-12-02T17:28:53,253 DEBUG [MemStoreFlusher.0 {}] regionserver.FlushAllLargeStoresPolicy(69): Since none of the CFs were above the size, flushing all. 2024-12-02T17:28:53,253 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=5.11 KB heapSize=8.96 KB 2024-12-02T17:28:53,253 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-15-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=13, ppid=10, state=RUNNABLE, hasLock=false; OpenRegionProcedure cd0c090bdce692f9dab734ec9b0dfe67, server=3b3375d1a925,40037,1733160507453 because future has completed 2024-12-02T17:28:53,253 INFO [StoreOpener-b10b696b6a97e7d9d531081df90f6c2a-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region b10b696b6a97e7d9d531081df90f6c2a columnFamilyName info 2024-12-02T17:28:53,253 DEBUG [StoreOpener-b10b696b6a97e7d9d531081df90f6c2a-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:28:53,257 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=13, resume processing ppid=10 2024-12-02T17:28:53,257 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=13, ppid=10, state=SUCCESS, hasLock=false; OpenRegionProcedure cd0c090bdce692f9dab734ec9b0dfe67, server=3b3375d1a925,40037,1733160507453 in 180 msec 2024-12-02T17:28:53,259 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=10, ppid=7, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=cd0c090bdce692f9dab734ec9b0dfe67, ASSIGN in 339 msec 2024-12-02T17:28:53,264 DEBUG [StoreOpener-b10b696b6a97e7d9d531081df90f6c2a-1 {}] regionserver.StoreEngine(278): loaded hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-c5f21f4850c9479aba2e5dc3dec7b9a2 2024-12-02T17:28:53,268 DEBUG [StoreOpener-b10b696b6a97e7d9d531081df90f6c2a-1 {}] regionserver.StoreEngine(278): loaded hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-e1c61ca3dbfd4c3fade405d03f737936 2024-12-02T17:28:53,272 DEBUG [StoreOpener-b10b696b6a97e7d9d531081df90f6c2a-1 {}] regionserver.StoreEngine(278): loaded hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5->hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935-top 2024-12-02T17:28:53,272 INFO [StoreOpener-b10b696b6a97e7d9d531081df90f6c2a-1 {}] regionserver.HStore(327): Store=b10b696b6a97e7d9d531081df90f6c2a/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:28:53,272 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(1038): replaying wal for b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,273 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,274 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/info/710948835bcc494da06bf8ccdc506d4a is 193, key is TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a./info:regioninfo/1733160533071/Put/seqid=0 2024-12-02T17:28:53,274 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,274 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(1048): stopping wal replay for b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,274 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(1060): Cleaning up temporary data for b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,276 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(1093): writing seq id for b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,277 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(1114): Opened b10b696b6a97e7d9d531081df90f6c2a; next sequenceid=122; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=797283, jitterRate=0.013798579573631287}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T17:28:53,277 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(1122): Running coprocessor post-open hooks for b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:28:53,278 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegion(1006): Region open journal for b10b696b6a97e7d9d531081df90f6c2a: Running coprocessor pre-open hook at 1733160533251Writing region info on filesystem at 1733160533251Initializing all the Stores at 1733160533252 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160533252Cleaning up temporary data from old regions at 1733160533274 (+22 ms)Running coprocessor post-open hooks at 1733160533277 (+3 ms)Region opened successfully at 1733160533277 2024-12-02T17:28:53,278 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegionServer(2236): Post open deploy tasks for TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., pid=12, masterSystemTime=1733160533225 2024-12-02T17:28:53,279 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.CompactSplit(403): Add compact mark for store b10b696b6a97e7d9d531081df90f6c2a:info, priority=-2147483648, current under compaction store size is 2 2024-12-02T17:28:53,279 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: Opening Region; compactionQueue=(longCompactions=0:shortCompactions=1), splitQueue=0 2024-12-02T17:28:53,279 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741850_1026 (size=9847) 2024-12-02T17:28:53,279 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:28:53,279 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): cd0c090bdce692f9dab734ec9b0dfe67#info#compaction#65 average throughput is 20.87 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:28:53,280 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741850_1026 (size=9847) 2024-12-02T17:28:53,280 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/.tmp/info/6f01375aea3b432a970bab0100a39523 is 1080, key is row0001/info:/1733160518306/Put/seqid=0 2024-12-02T17:28:53,280 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=4.92 KB at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/info/710948835bcc494da06bf8ccdc506d4a 2024-12-02T17:28:53,281 DEBUG [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] regionserver.HRegionServer(2266): Finished post open deploy task for TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:28:53,281 INFO [RS_OPEN_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_REGION, pid=12}] handler.AssignRegionHandler(153): Opened TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:28:53,281 INFO [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.HStore(1527): Keeping/Overriding Compaction request priority to -2147482648 for CF info since it belongs to recently split daughter region TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:28:53,281 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.HStore(1541): b10b696b6a97e7d9d531081df90f6c2a/info is initiating minor compaction (all files) 2024-12-02T17:28:53,281 INFO [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of b10b696b6a97e7d9d531081df90f6c2a/info in TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:28:53,281 INFO [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5->hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935-top, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-e1c61ca3dbfd4c3fade405d03f737936, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-c5f21f4850c9479aba2e5dc3dec7b9a2] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp, totalSize=113.9 K 2024-12-02T17:28:53,282 INFO [PEWorker-4 {}] assignment.RegionStateStore(223): pid=11 updating hbase:meta row=b10b696b6a97e7d9d531081df90f6c2a, regionState=OPEN, openSeqNum=122, regionLocation=3b3375d1a925,40037,1733160507453 2024-12-02T17:28:53,282 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] compactions.Compactor(225): Compacting c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5, keycount=32, bloomtype=ROW, size=73.6 K, encoding=NONE, compression=NONE, seqNum=83, earliestPutTs=1733160518306 2024-12-02T17:28:53,283 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] compactions.Compactor(225): Compacting TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-e1c61ca3dbfd4c3fade405d03f737936, keycount=16, bloomtype=ROW, size=21.7 K, encoding=NONE, compression=NONE, seqNum=101, earliestPutTs=1733160532523 2024-12-02T17:28:53,284 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] compactions.Compactor(225): Compacting TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-c5f21f4850c9479aba2e5dc3dec7b9a2, keycount=13, bloomtype=ROW, size=18.5 K, encoding=NONE, compression=NONE, seqNum=118, earliestPutTs=1733160532550 2024-12-02T17:28:53,284 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-15-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=12, ppid=11, state=RUNNABLE, hasLock=false; OpenRegionProcedure b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 because future has completed 2024-12-02T17:28:53,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741851_1027 (size=70862) 2024-12-02T17:28:53,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741851_1027 (size=70862) 2024-12-02T17:28:53,290 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=12, resume processing ppid=11 2024-12-02T17:28:53,290 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1521): Finished pid=12, ppid=11, state=SUCCESS, hasLock=false; OpenRegionProcedure b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 in 214 msec 2024-12-02T17:28:53,293 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/.tmp/info/6f01375aea3b432a970bab0100a39523 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/info/6f01375aea3b432a970bab0100a39523 2024-12-02T17:28:53,294 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=11, resume processing ppid=7 2024-12-02T17:28:53,294 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1521): Finished pid=11, ppid=7, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=b10b696b6a97e7d9d531081df90f6c2a, ASSIGN in 373 msec 2024-12-02T17:28:53,296 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=2bd92dcfc9073d608b4ecf1e0d6b55d5, daughterA=cd0c090bdce692f9dab734ec9b0dfe67, daughterB=b10b696b6a97e7d9d531081df90f6c2a in 697 msec 2024-12-02T17:28:53,300 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 1 (all) file(s) in cd0c090bdce692f9dab734ec9b0dfe67/info of cd0c090bdce692f9dab734ec9b0dfe67 into 6f01375aea3b432a970bab0100a39523(size=69.2 K), total size for store is 69.2 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:28:53,300 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for cd0c090bdce692f9dab734ec9b0dfe67: 2024-12-02T17:28:53,300 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67., storeName=cd0c090bdce692f9dab734ec9b0dfe67/info, priority=15, startTime=1733160533248; duration=0sec 2024-12-02T17:28:53,300 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:28:53,300 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: cd0c090bdce692f9dab734ec9b0dfe67:info 2024-12-02T17:28:53,305 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/ns/34c79883c30d419d8677ab7cd6a8f06f is 43, key is default/ns:d/1733160508229/Put/seqid=0 2024-12-02T17:28:53,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741852_1028 (size=5153) 2024-12-02T17:28:53,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741852_1028 (size=5153) 2024-12-02T17:28:53,312 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/ns/34c79883c30d419d8677ab7cd6a8f06f 2024-12-02T17:28:53,317 INFO [RS:0;3b3375d1a925:40037-longCompactions-0 {}] throttle.PressureAwareThroughputController(145): b10b696b6a97e7d9d531081df90f6c2a#info#compaction#67 average throughput is 33.86 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:28:53,317 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/c512335dc2ea4f7c8e835951dbde5f1b is 1080, key is row0062/info:/1733160530516/Put/seqid=0 2024-12-02T17:28:53,325 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741853_1029 (size=40830) 2024-12-02T17:28:53,325 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741853_1029 (size=40830) 2024-12-02T17:28:53,330 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/c512335dc2ea4f7c8e835951dbde5f1b as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c512335dc2ea4f7c8e835951dbde5f1b 2024-12-02T17:28:53,335 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/table/144998c6b38f4d4b8902f1c13100895b is 65, key is TestLogRolling-testLogRolling/table:state/1733160508638/Put/seqid=0 2024-12-02T17:28:53,336 INFO [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in b10b696b6a97e7d9d531081df90f6c2a/info of b10b696b6a97e7d9d531081df90f6c2a into c512335dc2ea4f7c8e835951dbde5f1b(size=39.9 K), total size for store is 39.9 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:28:53,336 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:28:53,336 INFO [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., storeName=b10b696b6a97e7d9d531081df90f6c2a/info, priority=13, startTime=1733160533278; duration=0sec 2024-12-02T17:28:53,336 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:28:53,336 DEBUG [RS:0;3b3375d1a925:40037-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: b10b696b6a97e7d9d531081df90f6c2a:info 2024-12-02T17:28:53,339 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741854_1030 (size=5340) 2024-12-02T17:28:53,339 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741854_1030 (size=5340) 2024-12-02T17:28:53,340 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=122 B at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/table/144998c6b38f4d4b8902f1c13100895b 2024-12-02T17:28:53,345 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/info/710948835bcc494da06bf8ccdc506d4a as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/info/710948835bcc494da06bf8ccdc506d4a 2024-12-02T17:28:53,349 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/info/710948835bcc494da06bf8ccdc506d4a, entries=30, sequenceid=17, filesize=9.6 K 2024-12-02T17:28:53,350 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/ns/34c79883c30d419d8677ab7cd6a8f06f as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/ns/34c79883c30d419d8677ab7cd6a8f06f 2024-12-02T17:28:53,354 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/ns/34c79883c30d419d8677ab7cd6a8f06f, entries=2, sequenceid=17, filesize=5.0 K 2024-12-02T17:28:53,355 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/table/144998c6b38f4d4b8902f1c13100895b as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/table/144998c6b38f4d4b8902f1c13100895b 2024-12-02T17:28:53,359 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/table/144998c6b38f4d4b8902f1c13100895b, entries=2, sequenceid=17, filesize=5.2 K 2024-12-02T17:28:53,360 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~5.11 KB/5234, heapSize ~8.66 KB/8872, currentSize=705 B/705 for 1588230740 in 107ms, sequenceid=17, compaction requested=false 2024-12-02T17:28:53,360 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for 1588230740: 2024-12-02T17:28:57,393 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:28:57,810 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,810 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,810 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,810 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,810 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,811 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,811 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,811 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,827 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,827 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,827 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,827 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,828 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,828 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,830 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,830 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,831 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:57,833 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,339 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:28:58,340 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,340 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,340 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,341 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,341 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,341 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,341 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,342 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,358 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,359 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,359 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,359 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,360 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,360 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,363 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,363 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,364 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:28:58,366 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:02,610 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] ipc.CallRunner(138): callId: 102 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.3:51472 deadline: 1733160552609, exception=org.apache.hadoop.hbase.NotServingRegionException: TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. is not online on 3b3375d1a925,40037,1733160507453 2024-12-02T17:29:02,610 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(64): Try updating region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 , the old value is region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2, error=org.apache.hadoop.hbase.NotServingRegionException: org.apache.hadoop.hbase.NotServingRegionException: TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. is not online on 3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:3186) at org.apache.hadoop.hbase.regionserver.HRegionServer.getRegion(HRegionServer.java:3164) at org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1413) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2943) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:29:02,610 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(72): The actual exception when updating region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 is org.apache.hadoop.hbase.NotServingRegionException: org.apache.hadoop.hbase.NotServingRegionException: TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5. is not online on 3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:3186) at org.apache.hadoop.hbase.regionserver.HRegionServer.getRegion(HRegionServer.java:3164) at org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1413) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2943) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:29:02,611 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(88): Try removing region=TestLogRolling-testLogRolling,,1733160508279.2bd92dcfc9073d608b4ecf1e0d6b55d5., hostname=3b3375d1a925,40037,1733160507453, seqNum=2 from cache 2024-12-02T17:29:08,333 INFO [master/3b3375d1a925:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T17:29:08,333 INFO [master/3b3375d1a925:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T17:29:13,198 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(329): Region 1588230740, had cached 0 bytes from a total of 20340 2024-12-02T17:29:22,774 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestLogRolling-testLogRolling', row='row0095', locateType=CURRENT is [region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122] 2024-12-02T17:29:24,786 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:24,786 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:29:24,791 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/9683618532a645f58c4f94f570809693 is 1080, key is row0095/info:/1733160562775/Put/seqid=0 2024-12-02T17:29:24,795 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741855_1031 (size=12513) 2024-12-02T17:29:24,796 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741855_1031 (size=12513) 2024-12-02T17:29:24,796 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=132 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/9683618532a645f58c4f94f570809693 2024-12-02T17:29:24,802 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/9683618532a645f58c4f94f570809693 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9683618532a645f58c4f94f570809693 2024-12-02T17:29:24,807 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9683618532a645f58c4f94f570809693, entries=7, sequenceid=132, filesize=12.2 K 2024-12-02T17:29:24,809 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=14.71 KB/15064 for b10b696b6a97e7d9d531081df90f6c2a in 23ms, sequenceid=132, compaction requested=false 2024-12-02T17:29:24,809 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:24,809 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:24,809 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=15.76 KB heapSize=17.13 KB 2024-12-02T17:29:24,813 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/0ab73d128f184c179207c7280ae1e0c6 is 1080, key is row0102/info:/1733160564787/Put/seqid=0 2024-12-02T17:29:24,817 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741856_1032 (size=21156) 2024-12-02T17:29:24,818 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741856_1032 (size=21156) 2024-12-02T17:29:24,818 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=15.76 KB at sequenceid=150 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/0ab73d128f184c179207c7280ae1e0c6 2024-12-02T17:29:24,823 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/0ab73d128f184c179207c7280ae1e0c6 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0ab73d128f184c179207c7280ae1e0c6 2024-12-02T17:29:24,847 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0ab73d128f184c179207c7280ae1e0c6, entries=15, sequenceid=150, filesize=20.7 K 2024-12-02T17:29:24,848 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~15.76 KB/16140, heapSize ~17.11 KB/17520, currentSize=12.61 KB/12912 for b10b696b6a97e7d9d531081df90f6c2a in 39ms, sequenceid=150, compaction requested=true 2024-12-02T17:29:24,848 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:24,848 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store b10b696b6a97e7d9d531081df90f6c2a:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:29:24,848 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:29:24,848 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:29:24,849 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 74499 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:29:24,849 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): b10b696b6a97e7d9d531081df90f6c2a/info is initiating minor compaction (all files) 2024-12-02T17:29:24,849 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of b10b696b6a97e7d9d531081df90f6c2a/info in TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:29:24,849 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c512335dc2ea4f7c8e835951dbde5f1b, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9683618532a645f58c4f94f570809693, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0ab73d128f184c179207c7280ae1e0c6] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp, totalSize=72.8 K 2024-12-02T17:29:24,850 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting c512335dc2ea4f7c8e835951dbde5f1b, keycount=33, bloomtype=ROW, size=39.9 K, encoding=NONE, compression=NONE, seqNum=118, earliestPutTs=1733160530516 2024-12-02T17:29:24,850 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 9683618532a645f58c4f94f570809693, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=132, earliestPutTs=1733160562775 2024-12-02T17:29:24,850 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 0ab73d128f184c179207c7280ae1e0c6, keycount=15, bloomtype=ROW, size=20.7 K, encoding=NONE, compression=NONE, seqNum=150, earliestPutTs=1733160564787 2024-12-02T17:29:24,863 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): b10b696b6a97e7d9d531081df90f6c2a#info#compaction#71 average throughput is 28.22 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:29:24,863 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/f2b2008a298a4aa4a9c061649ce461f4 is 1080, key is row0062/info:/1733160530516/Put/seqid=0 2024-12-02T17:29:24,877 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741857_1033 (size=64713) 2024-12-02T17:29:24,877 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741857_1033 (size=64713) 2024-12-02T17:29:24,883 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/f2b2008a298a4aa4a9c061649ce461f4 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/f2b2008a298a4aa4a9c061649ce461f4 2024-12-02T17:29:24,889 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in b10b696b6a97e7d9d531081df90f6c2a/info of b10b696b6a97e7d9d531081df90f6c2a into f2b2008a298a4aa4a9c061649ce461f4(size=63.2 K), total size for store is 63.2 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:29:24,889 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:24,889 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., storeName=b10b696b6a97e7d9d531081df90f6c2a/info, priority=13, startTime=1733160564848; duration=0sec 2024-12-02T17:29:24,889 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:29:24,889 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: b10b696b6a97e7d9d531081df90f6c2a:info 2024-12-02T17:29:26,849 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:26,849 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=13.66 KB heapSize=14.88 KB 2024-12-02T17:29:26,853 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/5713709879c74859a9c4a8d575b55cc8 is 1080, key is row0117/info:/1733160564810/Put/seqid=0 2024-12-02T17:29:26,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741858_1034 (size=19000) 2024-12-02T17:29:26,862 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741858_1034 (size=19000) 2024-12-02T17:29:26,873 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(5310): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:3.0.0-beta-2-SNAPSHOT] 2024-12-02T17:29:26,873 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] ipc.CallRunner(138): callId: 156 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.3:51472 deadline: 1733160576872, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 2024-12-02T17:29:26,873 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(64): Try updating region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 , the old value is region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122, error=org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:29:26,873 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(72): The actual exception when updating region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 is org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:29:26,873 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(76): Will not update region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 because the exception is null or not the one we care about 2024-12-02T17:29:27,260 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=13.66 KB at sequenceid=167 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/5713709879c74859a9c4a8d575b55cc8 2024-12-02T17:29:27,266 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/5713709879c74859a9c4a8d575b55cc8 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/5713709879c74859a9c4a8d575b55cc8 2024-12-02T17:29:27,271 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/5713709879c74859a9c4a8d575b55cc8, entries=13, sequenceid=167, filesize=18.6 K 2024-12-02T17:29:27,272 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~13.66 KB/13988, heapSize ~14.86 KB/15216, currentSize=16.81 KB/17216 for b10b696b6a97e7d9d531081df90f6c2a in 422ms, sequenceid=167, compaction requested=false 2024-12-02T17:29:27,272 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:27,393 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:29:28,237 DEBUG [master/3b3375d1a925:0.Chore.1 {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-02T17:29:36,950 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:36,950 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=17.86 KB heapSize=19.38 KB 2024-12-02T17:29:36,955 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/31cf893a27e944edae7a72a3698415dc is 1080, key is row0130/info:/1733160566850/Put/seqid=0 2024-12-02T17:29:36,970 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(5310): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:3.0.0-beta-2-SNAPSHOT] 2024-12-02T17:29:36,971 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] ipc.CallRunner(138): callId: 170 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.3:51472 deadline: 1733160586970, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 2024-12-02T17:29:36,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741859_1035 (size=23316) 2024-12-02T17:29:36,971 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(64): Try updating region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 , the old value is region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122, error=org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:29:36,971 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(72): The actual exception when updating region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 is org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:29:36,972 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(76): Will not update region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 because the exception is null or not the one we care about 2024-12-02T17:29:36,972 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741859_1035 (size=23316) 2024-12-02T17:29:36,972 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=17.86 KB at sequenceid=187 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/31cf893a27e944edae7a72a3698415dc 2024-12-02T17:29:36,979 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/31cf893a27e944edae7a72a3698415dc as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/31cf893a27e944edae7a72a3698415dc 2024-12-02T17:29:36,983 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/31cf893a27e944edae7a72a3698415dc, entries=17, sequenceid=187, filesize=22.8 K 2024-12-02T17:29:36,984 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~17.86 KB/18292, heapSize ~19.36 KB/19824, currentSize=12.61 KB/12912 for b10b696b6a97e7d9d531081df90f6c2a in 34ms, sequenceid=187, compaction requested=true 2024-12-02T17:29:36,984 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:36,984 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store b10b696b6a97e7d9d531081df90f6c2a:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:29:36,984 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:29:36,984 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:29:36,985 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 107029 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:29:36,985 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): b10b696b6a97e7d9d531081df90f6c2a/info is initiating minor compaction (all files) 2024-12-02T17:29:36,985 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of b10b696b6a97e7d9d531081df90f6c2a/info in TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:29:36,985 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/f2b2008a298a4aa4a9c061649ce461f4, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/5713709879c74859a9c4a8d575b55cc8, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/31cf893a27e944edae7a72a3698415dc] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp, totalSize=104.5 K 2024-12-02T17:29:36,985 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting f2b2008a298a4aa4a9c061649ce461f4, keycount=55, bloomtype=ROW, size=63.2 K, encoding=NONE, compression=NONE, seqNum=150, earliestPutTs=1733160530516 2024-12-02T17:29:36,986 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 5713709879c74859a9c4a8d575b55cc8, keycount=13, bloomtype=ROW, size=18.6 K, encoding=NONE, compression=NONE, seqNum=167, earliestPutTs=1733160564810 2024-12-02T17:29:36,986 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 31cf893a27e944edae7a72a3698415dc, keycount=17, bloomtype=ROW, size=22.8 K, encoding=NONE, compression=NONE, seqNum=187, earliestPutTs=1733160566850 2024-12-02T17:29:36,996 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): b10b696b6a97e7d9d531081df90f6c2a#info#compaction#74 average throughput is 43.61 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:29:36,996 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/2ddc657f740644658a6a69cd75b5d5d8 is 1080, key is row0062/info:/1733160530516/Put/seqid=0 2024-12-02T17:29:37,001 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741860_1036 (size=97232) 2024-12-02T17:29:37,002 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741860_1036 (size=97232) 2024-12-02T17:29:37,007 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/2ddc657f740644658a6a69cd75b5d5d8 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/2ddc657f740644658a6a69cd75b5d5d8 2024-12-02T17:29:37,013 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in b10b696b6a97e7d9d531081df90f6c2a/info of b10b696b6a97e7d9d531081df90f6c2a into 2ddc657f740644658a6a69cd75b5d5d8(size=95.0 K), total size for store is 95.0 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:29:37,013 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:37,013 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., storeName=b10b696b6a97e7d9d531081df90f6c2a/info, priority=13, startTime=1733160576984; duration=0sec 2024-12-02T17:29:37,013 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:29:37,013 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: b10b696b6a97e7d9d531081df90f6c2a:info 2024-12-02T17:29:38,229 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(329): Region cd0c090bdce692f9dab734ec9b0dfe67, had cached 0 bytes from a total of 70862 2024-12-02T17:29:38,251 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(329): Region b10b696b6a97e7d9d531081df90f6c2a, had cached 0 bytes from a total of 97232 2024-12-02T17:29:38,305 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): data stats (chunk size=2097152): current pool size=3, created chunk count=9, reused chunk count=62, reuseRatio=87.32% 2024-12-02T17:29:38,305 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): index stats (chunk size=209715): current pool size=0, created chunk count=0, reused chunk count=0, reuseRatio=0 2024-12-02T17:29:45,439 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,440 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,440 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,440 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,440 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,440 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,441 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,441 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,456 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,456 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,457 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,457 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,457 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,457 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,460 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,460 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,460 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,462 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,968 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T17:29:45,969 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,969 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,969 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,970 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,970 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,970 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,970 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,970 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,986 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,987 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,987 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,987 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,987 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,987 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,990 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,990 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,991 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:45,993 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T17:29:47,000 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:47,001 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=13.66 KB heapSize=14.88 KB 2024-12-02T17:29:47,005 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/0cd8e1d6159b47e9899220820f496a8e is 1080, key is row0147/info:/1733160576952/Put/seqid=0 2024-12-02T17:29:47,009 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741861_1037 (size=19000) 2024-12-02T17:29:47,010 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741861_1037 (size=19000) 2024-12-02T17:29:47,010 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=13.66 KB at sequenceid=204 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/0cd8e1d6159b47e9899220820f496a8e 2024-12-02T17:29:47,016 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/0cd8e1d6159b47e9899220820f496a8e as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0cd8e1d6159b47e9899220820f496a8e 2024-12-02T17:29:47,021 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0cd8e1d6159b47e9899220820f496a8e, entries=13, sequenceid=204, filesize=18.6 K 2024-12-02T17:29:47,022 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~13.66 KB/13988, heapSize ~14.86 KB/15216, currentSize=1.05 KB/1076 for b10b696b6a97e7d9d531081df90f6c2a in 21ms, sequenceid=204, compaction requested=false 2024-12-02T17:29:47,022 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:49,012 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:49,012 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:29:49,016 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/262546bbc93e406db66bf7cdec702b62 is 1080, key is row0160/info:/1733160587002/Put/seqid=0 2024-12-02T17:29:49,021 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741862_1038 (size=12516) 2024-12-02T17:29:49,022 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741862_1038 (size=12516) 2024-12-02T17:29:49,022 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=214 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/262546bbc93e406db66bf7cdec702b62 2024-12-02T17:29:49,028 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/262546bbc93e406db66bf7cdec702b62 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/262546bbc93e406db66bf7cdec702b62 2024-12-02T17:29:49,033 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/262546bbc93e406db66bf7cdec702b62, entries=7, sequenceid=214, filesize=12.2 K 2024-12-02T17:29:49,033 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=13.66 KB/13988 for b10b696b6a97e7d9d531081df90f6c2a in 21ms, sequenceid=214, compaction requested=true 2024-12-02T17:29:49,034 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:49,034 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store b10b696b6a97e7d9d531081df90f6c2a:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:29:49,034 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:29:49,034 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:29:49,035 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:49,035 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 128748 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:29:49,035 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=14.71 KB heapSize=16 KB 2024-12-02T17:29:49,035 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): b10b696b6a97e7d9d531081df90f6c2a/info is initiating minor compaction (all files) 2024-12-02T17:29:49,035 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of b10b696b6a97e7d9d531081df90f6c2a/info in TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:29:49,035 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/2ddc657f740644658a6a69cd75b5d5d8, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0cd8e1d6159b47e9899220820f496a8e, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/262546bbc93e406db66bf7cdec702b62] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp, totalSize=125.7 K 2024-12-02T17:29:49,035 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 2ddc657f740644658a6a69cd75b5d5d8, keycount=85, bloomtype=ROW, size=95.0 K, encoding=NONE, compression=NONE, seqNum=187, earliestPutTs=1733160530516 2024-12-02T17:29:49,036 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 0cd8e1d6159b47e9899220820f496a8e, keycount=13, bloomtype=ROW, size=18.6 K, encoding=NONE, compression=NONE, seqNum=204, earliestPutTs=1733160576952 2024-12-02T17:29:49,036 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 262546bbc93e406db66bf7cdec702b62, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=214, earliestPutTs=1733160587002 2024-12-02T17:29:49,039 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/1daf076b9f334286a7f386383cacad06 is 1080, key is row0167/info:/1733160589013/Put/seqid=0 2024-12-02T17:29:49,048 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741863_1039 (size=20078) 2024-12-02T17:29:49,049 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741863_1039 (size=20078) 2024-12-02T17:29:49,049 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): b10b696b6a97e7d9d531081df90f6c2a#info#compaction#78 average throughput is 53.87 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:29:49,049 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/a8f08a239d614aa68b4ce4be70f5c818 is 1080, key is row0062/info:/1733160530516/Put/seqid=0 2024-12-02T17:29:49,051 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=14.71 KB at sequenceid=231 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/1daf076b9f334286a7f386383cacad06 2024-12-02T17:29:49,056 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/1daf076b9f334286a7f386383cacad06 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/1daf076b9f334286a7f386383cacad06 2024-12-02T17:29:49,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741864_1040 (size=118898) 2024-12-02T17:29:49,061 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741864_1040 (size=118898) 2024-12-02T17:29:49,062 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/1daf076b9f334286a7f386383cacad06, entries=14, sequenceid=231, filesize=19.6 K 2024-12-02T17:29:49,063 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~14.71 KB/15064, heapSize ~15.98 KB/16368, currentSize=12.61 KB/12912 for b10b696b6a97e7d9d531081df90f6c2a in 28ms, sequenceid=231, compaction requested=false 2024-12-02T17:29:49,063 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:49,067 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/a8f08a239d614aa68b4ce4be70f5c818 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/a8f08a239d614aa68b4ce4be70f5c818 2024-12-02T17:29:49,072 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in b10b696b6a97e7d9d531081df90f6c2a/info of b10b696b6a97e7d9d531081df90f6c2a into a8f08a239d614aa68b4ce4be70f5c818(size=116.1 K), total size for store is 135.7 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:29:49,072 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:49,072 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., storeName=b10b696b6a97e7d9d531081df90f6c2a/info, priority=13, startTime=1733160589034; duration=0sec 2024-12-02T17:29:49,072 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:29:49,072 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: b10b696b6a97e7d9d531081df90f6c2a:info 2024-12-02T17:29:51,064 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:51,064 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=13.66 KB heapSize=14.88 KB 2024-12-02T17:29:51,069 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b is 1080, key is row0181/info:/1733160589035/Put/seqid=0 2024-12-02T17:29:51,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741865_1041 (size=19000) 2024-12-02T17:29:51,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741865_1041 (size=19000) 2024-12-02T17:29:51,076 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=13.66 KB at sequenceid=248 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b 2024-12-02T17:29:51,081 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b 2024-12-02T17:29:51,086 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b, entries=13, sequenceid=248, filesize=18.6 K 2024-12-02T17:29:51,087 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~13.66 KB/13988, heapSize ~14.86 KB/15216, currentSize=14.71 KB/15064 for b10b696b6a97e7d9d531081df90f6c2a in 22ms, sequenceid=248, compaction requested=true 2024-12-02T17:29:51,087 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:51,087 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store b10b696b6a97e7d9d531081df90f6c2a:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:29:51,087 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:29:51,087 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:29:51,088 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 157976 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:29:51,088 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:29:51,088 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): b10b696b6a97e7d9d531081df90f6c2a/info is initiating minor compaction (all files) 2024-12-02T17:29:51,088 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of b10b696b6a97e7d9d531081df90f6c2a/info in TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:29:51,088 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=15.76 KB heapSize=17.13 KB 2024-12-02T17:29:51,088 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/a8f08a239d614aa68b4ce4be70f5c818, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/1daf076b9f334286a7f386383cacad06, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp, totalSize=154.3 K 2024-12-02T17:29:51,089 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting a8f08a239d614aa68b4ce4be70f5c818, keycount=105, bloomtype=ROW, size=116.1 K, encoding=NONE, compression=NONE, seqNum=214, earliestPutTs=1733160530516 2024-12-02T17:29:51,089 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 1daf076b9f334286a7f386383cacad06, keycount=14, bloomtype=ROW, size=19.6 K, encoding=NONE, compression=NONE, seqNum=231, earliestPutTs=1733160589013 2024-12-02T17:29:51,089 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting c5127b9b4eb94c8e99b6b4ca8ad10e0b, keycount=13, bloomtype=ROW, size=18.6 K, encoding=NONE, compression=NONE, seqNum=248, earliestPutTs=1733160589035 2024-12-02T17:29:51,092 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/26863474a4274b919b380be7409b7c78 is 1080, key is row0194/info:/1733160591065/Put/seqid=0 2024-12-02T17:29:51,110 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(5310): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:3.0.0-beta-2-SNAPSHOT] 2024-12-02T17:29:51,110 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] ipc.CallRunner(138): callId: 235 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.3:51472 deadline: 1733160601110, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 2024-12-02T17:29:51,111 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(64): Try updating region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 , the old value is region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122, error=org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:29:51,111 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(72): The actual exception when updating region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 is org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:29:51,111 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(76): Will not update region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 because the exception is null or not the one we care about 2024-12-02T17:29:51,112 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741866_1042 (size=21165) 2024-12-02T17:29:51,113 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741866_1042 (size=21165) 2024-12-02T17:29:51,113 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=15.76 KB at sequenceid=266 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/26863474a4274b919b380be7409b7c78 2024-12-02T17:29:51,118 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): b10b696b6a97e7d9d531081df90f6c2a#info#compaction#81 average throughput is 45.15 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:29:51,119 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/26863474a4274b919b380be7409b7c78 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/26863474a4274b919b380be7409b7c78 2024-12-02T17:29:51,119 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/d5584797f86e4506a498c0cbfa624341 is 1080, key is row0062/info:/1733160530516/Put/seqid=0 2024-12-02T17:29:51,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741867_1043 (size=148311) 2024-12-02T17:29:51,123 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741867_1043 (size=148311) 2024-12-02T17:29:51,124 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/26863474a4274b919b380be7409b7c78, entries=15, sequenceid=266, filesize=20.7 K 2024-12-02T17:29:51,125 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~15.76 KB/16140, heapSize ~17.11 KB/17520, currentSize=14.71 KB/15064 for b10b696b6a97e7d9d531081df90f6c2a in 37ms, sequenceid=266, compaction requested=false 2024-12-02T17:29:51,125 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:51,128 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/d5584797f86e4506a498c0cbfa624341 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/d5584797f86e4506a498c0cbfa624341 2024-12-02T17:29:51,133 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in b10b696b6a97e7d9d531081df90f6c2a/info of b10b696b6a97e7d9d531081df90f6c2a into d5584797f86e4506a498c0cbfa624341(size=144.8 K), total size for store is 165.5 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:29:51,133 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:29:51,133 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., storeName=b10b696b6a97e7d9d531081df90f6c2a/info, priority=13, startTime=1733160591087; duration=0sec 2024-12-02T17:29:51,134 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:29:51,134 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: b10b696b6a97e7d9d531081df90f6c2a:info 2024-12-02T17:29:57,394 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtil$FsDatasetAsyncDiskServiceFixer(597): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T17:29:58,198 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(329): Region 1588230740, had cached 0 bytes from a total of 20340 2024-12-02T17:30:01,201 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:30:01,201 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=15.76 KB heapSize=17.13 KB 2024-12-02T17:30:01,205 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/23c46627fde9421ba0d3f575b1f7fb2c is 1080, key is row0209/info:/1733160591089/Put/seqid=0 2024-12-02T17:30:01,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741868_1044 (size=21171) 2024-12-02T17:30:01,211 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741868_1044 (size=21171) 2024-12-02T17:30:01,211 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=15.76 KB at sequenceid=285 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/23c46627fde9421ba0d3f575b1f7fb2c 2024-12-02T17:30:01,217 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/23c46627fde9421ba0d3f575b1f7fb2c as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/23c46627fde9421ba0d3f575b1f7fb2c 2024-12-02T17:30:01,221 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/23c46627fde9421ba0d3f575b1f7fb2c, entries=15, sequenceid=285, filesize=20.7 K 2024-12-02T17:30:01,222 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~15.76 KB/16140, heapSize ~17.11 KB/17520, currentSize=1.05 KB/1076 for b10b696b6a97e7d9d531081df90f6c2a in 21ms, sequenceid=285, compaction requested=true 2024-12-02T17:30:01,223 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:30:01,223 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store b10b696b6a97e7d9d531081df90f6c2a:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:30:01,223 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:30:01,223 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:30:01,224 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 190647 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:30:01,224 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): b10b696b6a97e7d9d531081df90f6c2a/info is initiating minor compaction (all files) 2024-12-02T17:30:01,224 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of b10b696b6a97e7d9d531081df90f6c2a/info in TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:30:01,224 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/d5584797f86e4506a498c0cbfa624341, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/26863474a4274b919b380be7409b7c78, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/23c46627fde9421ba0d3f575b1f7fb2c] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp, totalSize=186.2 K 2024-12-02T17:30:01,224 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting d5584797f86e4506a498c0cbfa624341, keycount=132, bloomtype=ROW, size=144.8 K, encoding=NONE, compression=NONE, seqNum=248, earliestPutTs=1733160530516 2024-12-02T17:30:01,225 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 26863474a4274b919b380be7409b7c78, keycount=15, bloomtype=ROW, size=20.7 K, encoding=NONE, compression=NONE, seqNum=266, earliestPutTs=1733160591065 2024-12-02T17:30:01,225 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 23c46627fde9421ba0d3f575b1f7fb2c, keycount=15, bloomtype=ROW, size=20.7 K, encoding=NONE, compression=NONE, seqNum=285, earliestPutTs=1733160591089 2024-12-02T17:30:01,237 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): b10b696b6a97e7d9d531081df90f6c2a#info#compaction#83 average throughput is 55.41 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:30:01,238 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/6c892963ab784a329d25666befb5b854 is 1080, key is row0062/info:/1733160530516/Put/seqid=0 2024-12-02T17:30:01,243 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741869_1045 (size=180785) 2024-12-02T17:30:01,243 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741869_1045 (size=180785) 2024-12-02T17:30:01,252 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/6c892963ab784a329d25666befb5b854 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6c892963ab784a329d25666befb5b854 2024-12-02T17:30:01,257 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in b10b696b6a97e7d9d531081df90f6c2a/info of b10b696b6a97e7d9d531081df90f6c2a into 6c892963ab784a329d25666befb5b854(size=176.5 K), total size for store is 176.5 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:30:01,257 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:30:01,257 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., storeName=b10b696b6a97e7d9d531081df90f6c2a/info, priority=13, startTime=1733160601223; duration=0sec 2024-12-02T17:30:01,257 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:30:01,257 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: b10b696b6a97e7d9d531081df90f6c2a:info 2024-12-02T17:30:03,212 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:30:03,213 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T17:30:03,218 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/6556715e97bb42a68c7d4e14baaa3b2e is 1080, key is row0224/info:/1733160601202/Put/seqid=0 2024-12-02T17:30:03,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741870_1046 (size=12523) 2024-12-02T17:30:03,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741870_1046 (size=12523) 2024-12-02T17:30:03,243 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(5310): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) ~[classes/:3.0.0-beta-2-SNAPSHOT] 2024-12-02T17:30:03,244 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] ipc.CallRunner(138): callId: 266 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.3:51472 deadline: 1733160613243, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 2024-12-02T17:30:03,244 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(64): Try updating region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 , the old value is region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122, error=org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:30:03,244 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(72): The actual exception when updating region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 is org.apache.hadoop.hbase.RegionTooBusyException: org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=b10b696b6a97e7d9d531081df90f6c2a, server=3b3375d1a925,40037,1733160507453 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5308) at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$11(HRegion.java:3329) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3322) at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3018) at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:2981) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:444) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102) at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82) 2024-12-02T17:30:03,244 DEBUG [RPCClient-NioEventLoopGroup-4-6 {}] client.AsyncRegionLocatorHelper(76): Will not update region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., hostname=3b3375d1a925,40037,1733160507453, seqNum=122 because the exception is null or not the one we care about 2024-12-02T17:30:03,630 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=296 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/6556715e97bb42a68c7d4e14baaa3b2e 2024-12-02T17:30:03,635 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/6556715e97bb42a68c7d4e14baaa3b2e as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6556715e97bb42a68c7d4e14baaa3b2e 2024-12-02T17:30:03,640 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6556715e97bb42a68c7d4e14baaa3b2e, entries=7, sequenceid=296, filesize=12.2 K 2024-12-02T17:30:03,641 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for b10b696b6a97e7d9d531081df90f6c2a in 428ms, sequenceid=296, compaction requested=false 2024-12-02T17:30:03,641 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:30:13,320 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40037 {}] regionserver.HRegion(8855): Flush requested on b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:30:13,320 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=24.17 KB heapSize=26.13 KB 2024-12-02T17:30:13,325 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/9f94787eed054bc5aec6c05687023c7f is 1080, key is row0231/info:/1733160603213/Put/seqid=0 2024-12-02T17:30:13,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741871_1047 (size=29807) 2024-12-02T17:30:13,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741871_1047 (size=29807) 2024-12-02T17:30:13,333 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.17 KB at sequenceid=322 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/9f94787eed054bc5aec6c05687023c7f 2024-12-02T17:30:13,338 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/9f94787eed054bc5aec6c05687023c7f as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9f94787eed054bc5aec6c05687023c7f 2024-12-02T17:30:13,342 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9f94787eed054bc5aec6c05687023c7f, entries=23, sequenceid=322, filesize=29.1 K 2024-12-02T17:30:13,343 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3140): Finished flush of dataSize ~24.17 KB/24748, heapSize ~26.11 KB/26736, currentSize=3.15 KB/3228 for b10b696b6a97e7d9d531081df90f6c2a in 23ms, sequenceid=322, compaction requested=true 2024-12-02T17:30:13,343 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:30:13,343 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store b10b696b6a97e7d9d531081df90f6c2a:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T17:30:13,343 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:30:13,343 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T17:30:13,344 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 223115 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T17:30:13,344 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1541): b10b696b6a97e7d9d531081df90f6c2a/info is initiating minor compaction (all files) 2024-12-02T17:30:13,344 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2416): Starting compaction of b10b696b6a97e7d9d531081df90f6c2a/info in TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:30:13,344 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6c892963ab784a329d25666befb5b854, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6556715e97bb42a68c7d4e14baaa3b2e, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9f94787eed054bc5aec6c05687023c7f] into tmpdir=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp, totalSize=217.9 K 2024-12-02T17:30:13,345 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 6c892963ab784a329d25666befb5b854, keycount=162, bloomtype=ROW, size=176.5 K, encoding=NONE, compression=NONE, seqNum=285, earliestPutTs=1733160530516 2024-12-02T17:30:13,345 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 6556715e97bb42a68c7d4e14baaa3b2e, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=296, earliestPutTs=1733160601202 2024-12-02T17:30:13,345 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] compactions.Compactor(225): Compacting 9f94787eed054bc5aec6c05687023c7f, keycount=23, bloomtype=ROW, size=29.1 K, encoding=NONE, compression=NONE, seqNum=322, earliestPutTs=1733160603213 2024-12-02T17:30:13,357 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): b10b696b6a97e7d9d531081df90f6c2a#info#compaction#86 average throughput is 65.67 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T17:30:13,358 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/566989aa3d7840079179a22f7ff62e5c is 1080, key is row0062/info:/1733160530516/Put/seqid=0 2024-12-02T17:30:13,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741872_1048 (size=213334) 2024-12-02T17:30:13,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741872_1048 (size=213334) 2024-12-02T17:30:13,365 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/566989aa3d7840079179a22f7ff62e5c as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/566989aa3d7840079179a22f7ff62e5c 2024-12-02T17:30:13,370 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HStore(1337): Completed compaction of 3 (all) file(s) in b10b696b6a97e7d9d531081df90f6c2a/info of b10b696b6a97e7d9d531081df90f6c2a into 566989aa3d7840079179a22f7ff62e5c(size=208.3 K), total size for store is 208.3 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T17:30:13,370 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.HRegion(2446): Compaction status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:30:13,370 INFO [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a., storeName=b10b696b6a97e7d9d531081df90f6c2a/info, priority=13, startTime=1733160613343; duration=0sec 2024-12-02T17:30:13,371 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T17:30:13,371 DEBUG [RS:0;3b3375d1a925:40037-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: b10b696b6a97e7d9d531081df90f6c2a:info 2024-12-02T17:30:15,326 INFO [Time-limited test {}] wal.AbstractTestLogRolling(285): after writing there are 0 log files 2024-12-02T17:30:15,326 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C40037%2C1733160507453.1733160615326 2024-12-02T17:30:15,342 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,342 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,342 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,342 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,342 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,342 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.1733160507829 with entries=308, filesize=307.10 KB; new WAL /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.1733160615326 2024-12-02T17:30:15,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741833_1009 (size=314476) 2024-12-02T17:30:15,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741833_1009 (size=314476) 2024-12-02T17:30:15,345 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40389:40389),(127.0.0.1/127.0.0.1:35929:35929)] 2024-12-02T17:30:15,348 DEBUG [Time-limited test {}] regionserver.HRegion(2603): Flush status journal for cd0c090bdce692f9dab734ec9b0dfe67: 2024-12-02T17:30:15,348 INFO [Time-limited test {}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=705 B heapSize=2.05 KB 2024-12-02T17:30:15,352 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/info/1419c7e6662840b5bece0399228b9eaa is 193, key is TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a./info:regioninfo/1733160533282/Put/seqid=0 2024-12-02T17:30:15,355 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741874_1050 (size=6223) 2024-12-02T17:30:15,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741874_1050 (size=6223) 2024-12-02T17:30:15,356 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=705 B at sequenceid=21 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/info/1419c7e6662840b5bece0399228b9eaa 2024-12-02T17:30:15,361 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/.tmp/info/1419c7e6662840b5bece0399228b9eaa as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/info/1419c7e6662840b5bece0399228b9eaa 2024-12-02T17:30:15,365 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/info/1419c7e6662840b5bece0399228b9eaa, entries=5, sequenceid=21, filesize=6.1 K 2024-12-02T17:30:15,366 INFO [Time-limited test {}] regionserver.HRegion(3140): Finished flush of dataSize ~705 B/705, heapSize ~1.29 KB/1320, currentSize=0 B/0 for 1588230740 in 18ms, sequenceid=21, compaction requested=false 2024-12-02T17:30:15,366 DEBUG [Time-limited test {}] regionserver.HRegion(2603): Flush status journal for 1588230740: 2024-12-02T17:30:15,366 INFO [Time-limited test {}] regionserver.HRegion(2902): Flushing b10b696b6a97e7d9d531081df90f6c2a 1/1 column families, dataSize=3.15 KB heapSize=3.63 KB 2024-12-02T17:30:15,369 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/a7269331c51b46939a6e4ac7d975e36c is 1080, key is row0254/info:/1733160613321/Put/seqid=0 2024-12-02T17:30:15,375 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741875_1051 (size=8199) 2024-12-02T17:30:15,375 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741875_1051 (size=8199) 2024-12-02T17:30:15,376 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=3.15 KB at sequenceid=329 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/a7269331c51b46939a6e4ac7d975e36c 2024-12-02T17:30:15,381 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/.tmp/info/a7269331c51b46939a6e4ac7d975e36c as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/a7269331c51b46939a6e4ac7d975e36c 2024-12-02T17:30:15,385 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/a7269331c51b46939a6e4ac7d975e36c, entries=3, sequenceid=329, filesize=8.0 K 2024-12-02T17:30:15,386 INFO [Time-limited test {}] regionserver.HRegion(3140): Finished flush of dataSize ~3.15 KB/3228, heapSize ~3.61 KB/3696, currentSize=0 B/0 for b10b696b6a97e7d9d531081df90f6c2a in 20ms, sequenceid=329, compaction requested=false 2024-12-02T17:30:15,386 DEBUG [Time-limited test {}] regionserver.HRegion(2603): Flush status journal for b10b696b6a97e7d9d531081df90f6c2a: 2024-12-02T17:30:15,386 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C40037%2C1733160507453.1733160615386 2024-12-02T17:30:15,393 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,393 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,393 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,393 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,393 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,394 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.1733160615326 with entries=2, filesize=723 B; new WAL /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.1733160615386 2024-12-02T17:30:15,394 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35929:35929),(127.0.0.1/127.0.0.1:40389:40389)] 2024-12-02T17:30:15,394 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.1733160615326 is not closed yet, will try archiving it next time 2024-12-02T17:30:15,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741873_1049 (size=731) 2024-12-02T17:30:15,395 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.1733160507829 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/oldWALs/3b3375d1a925%2C40037%2C1733160507453.1733160507829 2024-12-02T17:30:15,395 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [5,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T17:30:15,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741873_1049 (size=731) 2024-12-02T17:30:15,396 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/WALs/3b3375d1a925,40037,1733160507453/3b3375d1a925%2C40037%2C1733160507453.1733160615326 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/oldWALs/3b3375d1a925%2C40037%2C1733160507453.1733160615326 2024-12-02T17:30:15,495 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T17:30:15,495 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:30:15,496 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:30:15,496 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:15,496 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:15,496 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T17:30:15,496 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T17:30:15,496 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1124364620, stopped=false 2024-12-02T17:30:15,496 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=3b3375d1a925,39831,1733160507412 2024-12-02T17:30:15,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:30:15,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:30:15,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:15,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:15,498 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:30:15,498 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:30:15,498 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:30:15,498 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:15,499 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:30:15,499 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:30:15,499 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '3b3375d1a925,40037,1733160507453' ***** 2024-12-02T17:30:15,499 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T17:30:15,499 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T17:30:15,499 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T17:30:15,499 INFO [RS:0;3b3375d1a925:40037 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T17:30:15,499 INFO [RS:0;3b3375d1a925:40037 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T17:30:15,499 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(3091): Received CLOSE for cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:30:15,499 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(3091): Received CLOSE for b10b696b6a97e7d9d531081df90f6c2a 2024-12-02T17:30:15,499 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(959): stopping server 3b3375d1a925,40037,1733160507453 2024-12-02T17:30:15,500 INFO [RS:0;3b3375d1a925:40037 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:30:15,500 INFO [RS:0;3b3375d1a925:40037 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;3b3375d1a925:40037. 2024-12-02T17:30:15,500 DEBUG [RS:0;3b3375d1a925:40037 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:30:15,500 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing cd0c090bdce692f9dab734ec9b0dfe67, disabling compactions & flushes 2024-12-02T17:30:15,500 DEBUG [RS:0;3b3375d1a925:40037 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:15,500 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:30:15,500 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:30:15,500 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. after waiting 0 ms 2024-12-02T17:30:15,500 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T17:30:15,500 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:30:15,500 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T17:30:15,500 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T17:30:15,500 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T17:30:15,500 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1321): Waiting on 3 regions to close 2024-12-02T17:30:15,500 DEBUG [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1325): Online Regions={cd0c090bdce692f9dab734ec9b0dfe67=TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67., 1588230740=hbase:meta,,1.1588230740, b10b696b6a97e7d9d531081df90f6c2a=TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.} 2024-12-02T17:30:15,500 DEBUG [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1351): Waiting on 1588230740, b10b696b6a97e7d9d531081df90f6c2a, cd0c090bdce692f9dab734ec9b0dfe67 2024-12-02T17:30:15,500 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:30:15,500 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:30:15,500 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:30:15,500 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:30:15,500 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67.-1 {}] regionserver.HStore(2317): Moving the files [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5->hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935-bottom] to archive 2024-12-02T17:30:15,500 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:30:15,501 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T17:30:15,502 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:30:15,503 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67.-1 {}] regionserver.HRegionServer(3532): Failed to report file archival(s) to Master. This will be retried. org.apache.hadoop.hbase.ipc.StoppedRpcClientException: Call to address=3b3375d1a925:39831 failed on local exception: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.hbase.ipc.IPCUtil.wrapException(IPCUtil.java:237) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:395) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:430) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:425) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:117) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.setException(Call.java:132) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:451) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:336) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:606) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$BlockingStub.reportFileArchival(RegionServerStatusProtos.java:17350) ~[hbase-protocol-shaded-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegionServer.reportFileArchivalForQuotas(HRegionServer.java:3516) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.reportArchivedFilesForQuota(HStore.java:2412) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.removeCompactedfiles(HStore.java:2347) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.closeWithoutLock(HStore.java:738) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.close(HStore.java:804) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion$2.call(HRegion.java:1912) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion$2.call(HRegion.java:1909) ~[classes/:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) ~[?:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at org.apache.hadoop.hbase.ipc.AbstractRpcClient.getConnection(AbstractRpcClient.java:366) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:448) ~[hbase-client-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] ... 16 more 2024-12-02T17:30:15,503 WARN [StoreCloser-TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67.-1 {}] regionserver.HStore(2414): Failed to report archival of files: [] 2024-12-02T17:30:15,505 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/hbase/meta/1588230740/recovered.edits/24.seqid, newMaxSeqId=24, maxSeqId=1 2024-12-02T17:30:15,505 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:30:15,505 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:30:15,505 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160615500Running coprocessor pre-close hooks at 1733160615500Disabling compacts and flushes for region at 1733160615500Disabling writes for close at 1733160615500Writing region close event to WAL at 1733160615501 (+1 ms)Running coprocessor post-close hooks at 1733160615505 (+4 ms)Closed at 1733160615505 2024-12-02T17:30:15,505 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T17:30:15,506 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/cd0c090bdce692f9dab734ec9b0dfe67/recovered.edits/126.seqid, newMaxSeqId=126, maxSeqId=121 2024-12-02T17:30:15,507 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:30:15,507 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for cd0c090bdce692f9dab734ec9b0dfe67: Waiting for close lock at 1733160615500Running coprocessor pre-close hooks at 1733160615500Disabling compacts and flushes for region at 1733160615500Disabling writes for close at 1733160615500Writing region close event to WAL at 1733160615503 (+3 ms)Running coprocessor post-close hooks at 1733160615507 (+4 ms)Closed at 1733160615507 2024-12-02T17:30:15,507 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRolling,,1733160532597.cd0c090bdce692f9dab734ec9b0dfe67. 2024-12-02T17:30:15,507 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing b10b696b6a97e7d9d531081df90f6c2a, disabling compactions & flushes 2024-12-02T17:30:15,507 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:30:15,507 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:30:15,507 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. after waiting 0 ms 2024-12-02T17:30:15,507 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:30:15,507 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] regionserver.HStore(2317): Moving the files [hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5->hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/2bd92dcfc9073d608b4ecf1e0d6b55d5/info/c66a83493e5a4dfcada1bd023a2ae935-top, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-e1c61ca3dbfd4c3fade405d03f737936, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c512335dc2ea4f7c8e835951dbde5f1b, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-c5f21f4850c9479aba2e5dc3dec7b9a2, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9683618532a645f58c4f94f570809693, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/f2b2008a298a4aa4a9c061649ce461f4, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0ab73d128f184c179207c7280ae1e0c6, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/5713709879c74859a9c4a8d575b55cc8, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/2ddc657f740644658a6a69cd75b5d5d8, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/31cf893a27e944edae7a72a3698415dc, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0cd8e1d6159b47e9899220820f496a8e, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/a8f08a239d614aa68b4ce4be70f5c818, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/262546bbc93e406db66bf7cdec702b62, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/1daf076b9f334286a7f386383cacad06, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/d5584797f86e4506a498c0cbfa624341, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/26863474a4274b919b380be7409b7c78, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6c892963ab784a329d25666befb5b854, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/23c46627fde9421ba0d3f575b1f7fb2c, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6556715e97bb42a68c7d4e14baaa3b2e, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9f94787eed054bc5aec6c05687023c7f] to archive 2024-12-02T17:30:15,508 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T17:30:15,510 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c66a83493e5a4dfcada1bd023a2ae935.2bd92dcfc9073d608b4ecf1e0d6b55d5 2024-12-02T17:30:15,511 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-e1c61ca3dbfd4c3fade405d03f737936 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-e1c61ca3dbfd4c3fade405d03f737936 2024-12-02T17:30:15,512 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c512335dc2ea4f7c8e835951dbde5f1b to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c512335dc2ea4f7c8e835951dbde5f1b 2024-12-02T17:30:15,513 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-c5f21f4850c9479aba2e5dc3dec7b9a2 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/TestLogRolling-testLogRolling=2bd92dcfc9073d608b4ecf1e0d6b55d5-c5f21f4850c9479aba2e5dc3dec7b9a2 2024-12-02T17:30:15,514 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9683618532a645f58c4f94f570809693 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9683618532a645f58c4f94f570809693 2024-12-02T17:30:15,515 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/f2b2008a298a4aa4a9c061649ce461f4 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/f2b2008a298a4aa4a9c061649ce461f4 2024-12-02T17:30:15,516 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0ab73d128f184c179207c7280ae1e0c6 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0ab73d128f184c179207c7280ae1e0c6 2024-12-02T17:30:15,517 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/5713709879c74859a9c4a8d575b55cc8 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/5713709879c74859a9c4a8d575b55cc8 2024-12-02T17:30:15,519 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/2ddc657f740644658a6a69cd75b5d5d8 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/2ddc657f740644658a6a69cd75b5d5d8 2024-12-02T17:30:15,520 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/31cf893a27e944edae7a72a3698415dc to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/31cf893a27e944edae7a72a3698415dc 2024-12-02T17:30:15,521 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0cd8e1d6159b47e9899220820f496a8e to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/0cd8e1d6159b47e9899220820f496a8e 2024-12-02T17:30:15,522 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/a8f08a239d614aa68b4ce4be70f5c818 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/a8f08a239d614aa68b4ce4be70f5c818 2024-12-02T17:30:15,523 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/262546bbc93e406db66bf7cdec702b62 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/262546bbc93e406db66bf7cdec702b62 2024-12-02T17:30:15,524 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/1daf076b9f334286a7f386383cacad06 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/1daf076b9f334286a7f386383cacad06 2024-12-02T17:30:15,526 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/d5584797f86e4506a498c0cbfa624341 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/d5584797f86e4506a498c0cbfa624341 2024-12-02T17:30:15,527 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/c5127b9b4eb94c8e99b6b4ca8ad10e0b 2024-12-02T17:30:15,528 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/26863474a4274b919b380be7409b7c78 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/26863474a4274b919b380be7409b7c78 2024-12-02T17:30:15,529 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6c892963ab784a329d25666befb5b854 to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6c892963ab784a329d25666befb5b854 2024-12-02T17:30:15,530 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/23c46627fde9421ba0d3f575b1f7fb2c to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/23c46627fde9421ba0d3f575b1f7fb2c 2024-12-02T17:30:15,531 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6556715e97bb42a68c7d4e14baaa3b2e to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/6556715e97bb42a68c7d4e14baaa3b2e 2024-12-02T17:30:15,532 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9f94787eed054bc5aec6c05687023c7f to hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/archive/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/info/9f94787eed054bc5aec6c05687023c7f 2024-12-02T17:30:15,532 WARN [StoreCloser-TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a.-1 {}] regionserver.HStore(2414): Failed to report archival of files: [c512335dc2ea4f7c8e835951dbde5f1b=40830, 9683618532a645f58c4f94f570809693=12513, f2b2008a298a4aa4a9c061649ce461f4=64713, 0ab73d128f184c179207c7280ae1e0c6=21156, 5713709879c74859a9c4a8d575b55cc8=19000, 2ddc657f740644658a6a69cd75b5d5d8=97232, 31cf893a27e944edae7a72a3698415dc=23316, 0cd8e1d6159b47e9899220820f496a8e=19000, a8f08a239d614aa68b4ce4be70f5c818=118898, 262546bbc93e406db66bf7cdec702b62=12516, 1daf076b9f334286a7f386383cacad06=20078, d5584797f86e4506a498c0cbfa624341=148311, c5127b9b4eb94c8e99b6b4ca8ad10e0b=19000, 26863474a4274b919b380be7409b7c78=21165, 6c892963ab784a329d25666befb5b854=180785, 23c46627fde9421ba0d3f575b1f7fb2c=21171, 6556715e97bb42a68c7d4e14baaa3b2e=12523, 9f94787eed054bc5aec6c05687023c7f=29807] 2024-12-02T17:30:15,535 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/data/default/TestLogRolling-testLogRolling/b10b696b6a97e7d9d531081df90f6c2a/recovered.edits/332.seqid, newMaxSeqId=332, maxSeqId=121 2024-12-02T17:30:15,536 INFO [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:30:15,536 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for b10b696b6a97e7d9d531081df90f6c2a: Waiting for close lock at 1733160615507Running coprocessor pre-close hooks at 1733160615507Disabling compacts and flushes for region at 1733160615507Disabling writes for close at 1733160615507Writing region close event to WAL at 1733160615533 (+26 ms)Running coprocessor post-close hooks at 1733160615536 (+3 ms)Closed at 1733160615536 2024-12-02T17:30:15,536 DEBUG [RS_CLOSE_REGION-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRolling,row0062,1733160532597.b10b696b6a97e7d9d531081df90f6c2a. 2024-12-02T17:30:15,700 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(976): stopping server 3b3375d1a925,40037,1733160507453; all regions closed. 2024-12-02T17:30:15,701 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,701 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,701 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,701 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,701 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,702 INFO [regionserver/3b3375d1a925:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:30:15,703 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741834_1010 (size=8107) 2024-12-02T17:30:15,703 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741834_1010 (size=8107) 2024-12-02T17:30:15,705 DEBUG [RS:0;3b3375d1a925:40037 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/oldWALs 2024-12-02T17:30:15,705 INFO [RS:0;3b3375d1a925:40037 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C40037%2C1733160507453.meta:.meta(num 1733160508185) 2024-12-02T17:30:15,705 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,706 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,706 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,706 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,706 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,707 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741876_1052 (size=778) 2024-12-02T17:30:15,707 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741876_1052 (size=778) 2024-12-02T17:30:15,709 DEBUG [RS:0;3b3375d1a925:40037 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/oldWALs 2024-12-02T17:30:15,709 INFO [RS:0;3b3375d1a925:40037 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C40037%2C1733160507453:(num 1733160615386) 2024-12-02T17:30:15,709 DEBUG [RS:0;3b3375d1a925:40037 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:15,709 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:30:15,709 INFO [RS:0;3b3375d1a925:40037 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:30:15,709 INFO [RS:0;3b3375d1a925:40037 {}] hbase.ChoreService(370): Chore service for: regionserver/3b3375d1a925:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-02T17:30:15,709 INFO [RS:0;3b3375d1a925:40037 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:30:15,710 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:30:15,710 INFO [RS:0;3b3375d1a925:40037 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:40037 2024-12-02T17:30:15,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/3b3375d1a925,40037,1733160507453 2024-12-02T17:30:15,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:30:15,712 INFO [RS:0;3b3375d1a925:40037 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:30:15,713 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [3b3375d1a925,40037,1733160507453] 2024-12-02T17:30:15,714 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/3b3375d1a925,40037,1733160507453 already deleted, retry=false 2024-12-02T17:30:15,715 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 3b3375d1a925,40037,1733160507453 expired; onlineServers=0 2024-12-02T17:30:15,715 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '3b3375d1a925,39831,1733160507412' ***** 2024-12-02T17:30:15,715 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T17:30:15,715 INFO [M:0;3b3375d1a925:39831 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:30:15,715 INFO [M:0;3b3375d1a925:39831 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:30:15,715 DEBUG [M:0;3b3375d1a925:39831 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T17:30:15,715 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T17:30:15,715 DEBUG [M:0;3b3375d1a925:39831 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T17:30:15,715 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160507593 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160507593,5,FailOnTimeoutGroup] 2024-12-02T17:30:15,715 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160507592 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160507592,5,FailOnTimeoutGroup] 2024-12-02T17:30:15,715 INFO [M:0;3b3375d1a925:39831 {}] hbase.ChoreService(370): Chore service for: master/3b3375d1a925:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T17:30:15,715 INFO [M:0;3b3375d1a925:39831 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:30:15,715 DEBUG [M:0;3b3375d1a925:39831 {}] master.HMaster(1795): Stopping service threads 2024-12-02T17:30:15,715 INFO [M:0;3b3375d1a925:39831 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T17:30:15,715 INFO [M:0;3b3375d1a925:39831 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:30:15,715 INFO [M:0;3b3375d1a925:39831 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T17:30:15,715 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T17:30:15,717 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T17:30:15,717 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:15,717 DEBUG [M:0;3b3375d1a925:39831 {}] zookeeper.ZKUtil(347): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T17:30:15,717 WARN [M:0;3b3375d1a925:39831 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T17:30:15,717 INFO [M:0;3b3375d1a925:39831 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/.lastflushedseqids 2024-12-02T17:30:15,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741877_1053 (size=228) 2024-12-02T17:30:15,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741877_1053 (size=228) 2024-12-02T17:30:15,722 INFO [M:0;3b3375d1a925:39831 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T17:30:15,722 INFO [M:0;3b3375d1a925:39831 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T17:30:15,723 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:30:15,723 INFO [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:15,723 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:15,723 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:30:15,723 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:15,723 INFO [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=51.41 KB heapSize=63.33 KB 2024-12-02T17:30:15,738 DEBUG [M:0;3b3375d1a925:39831 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a849a04fc108452393a20ead5e3292c3 is 82, key is hbase:meta,,1/info:regioninfo/1733160508215/Put/seqid=0 2024-12-02T17:30:15,742 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741878_1054 (size=5672) 2024-12-02T17:30:15,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741878_1054 (size=5672) 2024-12-02T17:30:15,743 INFO [M:0;3b3375d1a925:39831 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=125 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a849a04fc108452393a20ead5e3292c3 2024-12-02T17:30:15,761 DEBUG [M:0;3b3375d1a925:39831 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/7a557042afa541d784d1fb9a168eca22 is 749, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733160508643/Put/seqid=0 2024-12-02T17:30:15,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741879_1055 (size=7089) 2024-12-02T17:30:15,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741879_1055 (size=7089) 2024-12-02T17:30:15,766 INFO [M:0;3b3375d1a925:39831 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=50.80 KB at sequenceid=125 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/7a557042afa541d784d1fb9a168eca22 2024-12-02T17:30:15,770 INFO [M:0;3b3375d1a925:39831 {}] regionserver.StoreFileReader(518): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 7a557042afa541d784d1fb9a168eca22 2024-12-02T17:30:15,783 DEBUG [M:0;3b3375d1a925:39831 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7a5bca25fb7d4bab89a795cfeee8e0e1 is 69, key is 3b3375d1a925,40037,1733160507453/rs:state/1733160507686/Put/seqid=0 2024-12-02T17:30:15,787 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741880_1056 (size=5156) 2024-12-02T17:30:15,787 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741880_1056 (size=5156) 2024-12-02T17:30:15,788 INFO [M:0;3b3375d1a925:39831 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=125 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7a5bca25fb7d4bab89a795cfeee8e0e1 2024-12-02T17:30:15,804 DEBUG [M:0;3b3375d1a925:39831 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/3c8ca0fde3e54aadbef7ae3a167f56de is 52, key is load_balancer_on/state:d/1733160508276/Put/seqid=0 2024-12-02T17:30:15,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741881_1057 (size=5056) 2024-12-02T17:30:15,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741881_1057 (size=5056) 2024-12-02T17:30:15,809 INFO [M:0;3b3375d1a925:39831 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=125 (bloomFilter=true), to=hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/3c8ca0fde3e54aadbef7ae3a167f56de 2024-12-02T17:30:15,813 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a849a04fc108452393a20ead5e3292c3 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/a849a04fc108452393a20ead5e3292c3 2024-12-02T17:30:15,813 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:30:15,813 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40037-0x1009c08a2a20001, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:30:15,813 INFO [RS:0;3b3375d1a925:40037 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:30:15,813 INFO [RS:0;3b3375d1a925:40037 {}] regionserver.HRegionServer(1031): Exiting; stopping=3b3375d1a925,40037,1733160507453; zookeeper connection closed. 2024-12-02T17:30:15,814 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@56c0e02c {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@56c0e02c 2024-12-02T17:30:15,814 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T17:30:15,817 INFO [M:0;3b3375d1a925:39831 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/a849a04fc108452393a20ead5e3292c3, entries=8, sequenceid=125, filesize=5.5 K 2024-12-02T17:30:15,818 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/7a557042afa541d784d1fb9a168eca22 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/7a557042afa541d784d1fb9a168eca22 2024-12-02T17:30:15,822 INFO [M:0;3b3375d1a925:39831 {}] regionserver.StoreFileReader(518): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 7a557042afa541d784d1fb9a168eca22 2024-12-02T17:30:15,822 INFO [M:0;3b3375d1a925:39831 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/7a557042afa541d784d1fb9a168eca22, entries=13, sequenceid=125, filesize=6.9 K 2024-12-02T17:30:15,822 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7a5bca25fb7d4bab89a795cfeee8e0e1 as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/7a5bca25fb7d4bab89a795cfeee8e0e1 2024-12-02T17:30:15,826 INFO [M:0;3b3375d1a925:39831 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/7a5bca25fb7d4bab89a795cfeee8e0e1, entries=1, sequenceid=125, filesize=5.0 K 2024-12-02T17:30:15,827 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/3c8ca0fde3e54aadbef7ae3a167f56de as hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/3c8ca0fde3e54aadbef7ae3a167f56de 2024-12-02T17:30:15,830 INFO [M:0;3b3375d1a925:39831 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42821/user/jenkins/test-data/c49de6a6-c0bb-d382-a18a-7184553012ec/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/3c8ca0fde3e54aadbef7ae3a167f56de, entries=1, sequenceid=125, filesize=4.9 K 2024-12-02T17:30:15,831 INFO [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(3140): Finished flush of dataSize ~51.41 KB/52639, heapSize ~63.27 KB/64784, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 108ms, sequenceid=125, compaction requested=false 2024-12-02T17:30:15,832 INFO [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:15,832 DEBUG [M:0;3b3375d1a925:39831 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160615723Disabling compacts and flushes for region at 1733160615723Disabling writes for close at 1733160615723Obtaining lock to block concurrent updates at 1733160615723Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733160615723Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=52639, getHeapSize=64784, getOffHeapSize=0, getCellsCount=148 at 1733160615723Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733160615724 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733160615724Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733160615738 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733160615738Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733160615747 (+9 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733160615760 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733160615760Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733160615770 (+10 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733160615783 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733160615783Flushing 1595e783b53d99cd5eef43b6debb2682/state: creating writer at 1733160615791 (+8 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: appending metadata at 1733160615804 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: closing flushed file at 1733160615804Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@735cd1b0: reopening flushed file at 1733160615812 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6e391bfd: reopening flushed file at 1733160615817 (+5 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@41d183b9: reopening flushed file at 1733160615822 (+5 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@275cc3e5: reopening flushed file at 1733160615826 (+4 ms)Finished flush of dataSize ~51.41 KB/52639, heapSize ~63.27 KB/64784, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 108ms, sequenceid=125, compaction requested=false at 1733160615831 (+5 ms)Writing region close event to WAL at 1733160615832 (+1 ms)Closed at 1733160615832 2024-12-02T17:30:15,833 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,833 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,833 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,833 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,833 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:15,835 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42947 is added to blk_1073741830_1006 (size=61308) 2024-12-02T17:30:15,835 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45313 is added to blk_1073741830_1006 (size=61308) 2024-12-02T17:30:15,836 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:30:15,836 INFO [M:0;3b3375d1a925:39831 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T17:30:15,836 INFO [M:0;3b3375d1a925:39831 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:39831 2024-12-02T17:30:15,837 INFO [M:0;3b3375d1a925:39831 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:30:15,939 INFO [M:0;3b3375d1a925:39831 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:30:15,939 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:30:15,939 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39831-0x1009c08a2a20000, quorum=127.0.0.1:53832, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:30:15,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@14646d9{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:30:15,942 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1d428ad5{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:30:15,942 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:30:15,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5b6783f5{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:30:15,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5fe51576{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.log.dir/,STOPPED} 2024-12-02T17:30:15,944 WARN [BP-436516782-172.17.0.3-1733160506756 heartbeating to localhost/127.0.0.1:42821 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:30:15,944 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:30:15,944 WARN [BP-436516782-172.17.0.3-1733160506756 heartbeating to localhost/127.0.0.1:42821 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-436516782-172.17.0.3-1733160506756 (Datanode Uuid d4cdc363-0de9-4f5d-9992-b88cee95e7a0) service to localhost/127.0.0.1:42821 2024-12-02T17:30:15,944 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:30:15,944 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/data/data3/current/BP-436516782-172.17.0.3-1733160506756 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:30:15,945 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/data/data4/current/BP-436516782-172.17.0.3-1733160506756 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:30:15,945 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:30:15,947 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@e1b4695{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:30:15,947 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1520cb76{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:30:15,947 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:30:15,947 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3aca12f3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:30:15,947 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3aa952ac{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.log.dir/,STOPPED} 2024-12-02T17:30:15,949 WARN [BP-436516782-172.17.0.3-1733160506756 heartbeating to localhost/127.0.0.1:42821 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:30:15,949 WARN [BP-436516782-172.17.0.3-1733160506756 heartbeating to localhost/127.0.0.1:42821 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-436516782-172.17.0.3-1733160506756 (Datanode Uuid ee7e4b05-cc51-4ee8-840f-45b80a5cd1c2) service to localhost/127.0.0.1:42821 2024-12-02T17:30:15,949 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:30:15,949 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:30:15,949 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/data/data1/current/BP-436516782-172.17.0.3-1733160506756 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:30:15,949 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/cluster_e2b7a787-d3c3-5186-d8a0-fa7bb344598f/data/data2/current/BP-436516782-172.17.0.3-1733160506756 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:30:15,950 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:30:15,955 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@40a5a9d2{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:30:15,956 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@655ba446{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:30:15,956 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:30:15,956 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@273a6f23{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:30:15,956 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2ffc76d1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.log.dir/,STOPPED} 2024-12-02T17:30:15,963 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T17:30:15,998 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T17:30:16,008 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRolling Thread=224 (was 203) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-15-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:42821 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-38-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-41-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-15-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-39-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-41-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42821 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-39-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-40-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.6@localhost:42821 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-14-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-40-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-15-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-41-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-38-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-14-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-14-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42821 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:42821 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-40-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-39-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-38-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:42821 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:42821 from jenkins.hfs.6 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42821 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=509 (was 483) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=137 (was 133) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=3105 (was 3184) 2024-12-02T17:30:16,016 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnNothingWritten Thread=224, OpenFileDescriptor=509, MaxFileDescriptor=1048576, SystemLoadAverage=137, ProcessCount=11, AvailableMemoryMB=3105 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.log.dir so I do NOT create it in target/test-data/0a023e16-09ff-b414-02d3-b38836200260 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/e8553c75-412c-a701-2f1a-97e6385a8b20/hadoop.tmp.dir so I do NOT create it in target/test-data/0a023e16-09ff-b414-02d3-b38836200260 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380, deleteOnExit=true 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/test.cache.data in system properties and HBase conf 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/hadoop.log.dir in system properties and HBase conf 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-02T17:30:16,017 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:30:16,017 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/nfs.dump.dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/java.io.tmpdir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T17:30:16,018 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T17:30:16,030 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:30:16,082 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:30:16,086 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:30:16,087 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:30:16,087 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:30:16,087 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:30:16,087 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:30:16,088 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3073e97e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:30:16,088 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4d118eec{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:30:16,202 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7b063aa1{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/java.io.tmpdir/jetty-localhost-33603-hadoop-hdfs-3_4_1-tests_jar-_-any-2396404875998003196/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:30:16,203 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@778227aa{HTTP/1.1, (http/1.1)}{localhost:33603} 2024-12-02T17:30:16,203 INFO [Time-limited test {}] server.Server(415): Started @343875ms 2024-12-02T17:30:16,215 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T17:30:16,264 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:30:16,266 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:30:16,267 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:30:16,267 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:30:16,267 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T17:30:16,268 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@18478920{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:30:16,268 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@40f3733a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:30:16,380 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@69a2ae1b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/java.io.tmpdir/jetty-localhost-43211-hadoop-hdfs-3_4_1-tests_jar-_-any-15747208690384745332/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:30:16,380 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@64e1b9c7{HTTP/1.1, (http/1.1)}{localhost:43211} 2024-12-02T17:30:16,380 INFO [Time-limited test {}] server.Server(415): Started @344053ms 2024-12-02T17:30:16,381 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:30:16,409 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T17:30:16,411 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T17:30:16,412 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T17:30:16,412 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T17:30:16,412 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T17:30:16,413 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2b13a29{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/hadoop.log.dir/,AVAILABLE} 2024-12-02T17:30:16,413 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@e9b8f9f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T17:30:16,456 WARN [Thread-2458 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/data/data2/current/BP-1495924061-172.17.0.3-1733160616036/current, will proceed with Du for space computation calculation, 2024-12-02T17:30:16,456 WARN [Thread-2457 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/data/data1/current/BP-1495924061-172.17.0.3-1733160616036/current, will proceed with Du for space computation calculation, 2024-12-02T17:30:16,472 WARN [Thread-2436 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:30:16,477 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xac13dd0c590e674d with lease ID 0x57fccf992d41e35b: Processing first storage report for DS-f5616bd1-1718-4da9-9f7e-ef3715f8265a from datanode DatanodeRegistration(127.0.0.1:33331, datanodeUuid=7effab17-28f7-4f2b-bef3-9d3e9a5fae23, infoPort=34697, infoSecurePort=0, ipcPort=37923, storageInfo=lv=-57;cid=testClusterID;nsid=939995934;c=1733160616036) 2024-12-02T17:30:16,477 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xac13dd0c590e674d with lease ID 0x57fccf992d41e35b: from storage DS-f5616bd1-1718-4da9-9f7e-ef3715f8265a node DatanodeRegistration(127.0.0.1:33331, datanodeUuid=7effab17-28f7-4f2b-bef3-9d3e9a5fae23, infoPort=34697, infoSecurePort=0, ipcPort=37923, storageInfo=lv=-57;cid=testClusterID;nsid=939995934;c=1733160616036), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:30:16,477 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xac13dd0c590e674d with lease ID 0x57fccf992d41e35b: Processing first storage report for DS-84828123-b124-4934-924e-c70b3a503c0c from datanode DatanodeRegistration(127.0.0.1:33331, datanodeUuid=7effab17-28f7-4f2b-bef3-9d3e9a5fae23, infoPort=34697, infoSecurePort=0, ipcPort=37923, storageInfo=lv=-57;cid=testClusterID;nsid=939995934;c=1733160616036) 2024-12-02T17:30:16,477 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xac13dd0c590e674d with lease ID 0x57fccf992d41e35b: from storage DS-84828123-b124-4934-924e-c70b3a503c0c node DatanodeRegistration(127.0.0.1:33331, datanodeUuid=7effab17-28f7-4f2b-bef3-9d3e9a5fae23, infoPort=34697, infoSecurePort=0, ipcPort=37923, storageInfo=lv=-57;cid=testClusterID;nsid=939995934;c=1733160616036), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:30:16,532 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@a8a2fb4{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/java.io.tmpdir/jetty-localhost-46477-hadoop-hdfs-3_4_1-tests_jar-_-any-10130041125807705584/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:30:16,532 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1f41372a{HTTP/1.1, (http/1.1)}{localhost:46477} 2024-12-02T17:30:16,532 INFO [Time-limited test {}] server.Server(415): Started @344204ms 2024-12-02T17:30:16,533 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T17:30:16,606 WARN [Thread-2483 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/data/data3/current/BP-1495924061-172.17.0.3-1733160616036/current, will proceed with Du for space computation calculation, 2024-12-02T17:30:16,606 WARN [Thread-2484 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/data/data4/current/BP-1495924061-172.17.0.3-1733160616036/current, will proceed with Du for space computation calculation, 2024-12-02T17:30:16,629 WARN [Thread-2472 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T17:30:16,630 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x16536762eb5ebba1 with lease ID 0x57fccf992d41e35c: Processing first storage report for DS-6c2d828a-1063-4164-ac90-345488819d98 from datanode DatanodeRegistration(127.0.0.1:41719, datanodeUuid=6ec327de-6496-47da-aa4f-b4f73202aa4f, infoPort=46875, infoSecurePort=0, ipcPort=34829, storageInfo=lv=-57;cid=testClusterID;nsid=939995934;c=1733160616036) 2024-12-02T17:30:16,631 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x16536762eb5ebba1 with lease ID 0x57fccf992d41e35c: from storage DS-6c2d828a-1063-4164-ac90-345488819d98 node DatanodeRegistration(127.0.0.1:41719, datanodeUuid=6ec327de-6496-47da-aa4f-b4f73202aa4f, infoPort=46875, infoSecurePort=0, ipcPort=34829, storageInfo=lv=-57;cid=testClusterID;nsid=939995934;c=1733160616036), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:30:16,631 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x16536762eb5ebba1 with lease ID 0x57fccf992d41e35c: Processing first storage report for DS-4074a4b4-6185-4121-bf0e-818e9b02d101 from datanode DatanodeRegistration(127.0.0.1:41719, datanodeUuid=6ec327de-6496-47da-aa4f-b4f73202aa4f, infoPort=46875, infoSecurePort=0, ipcPort=34829, storageInfo=lv=-57;cid=testClusterID;nsid=939995934;c=1733160616036) 2024-12-02T17:30:16,631 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x16536762eb5ebba1 with lease ID 0x57fccf992d41e35c: from storage DS-4074a4b4-6185-4121-bf0e-818e9b02d101 node DatanodeRegistration(127.0.0.1:41719, datanodeUuid=6ec327de-6496-47da-aa4f-b4f73202aa4f, infoPort=46875, infoSecurePort=0, ipcPort=34829, storageInfo=lv=-57;cid=testClusterID;nsid=939995934;c=1733160616036), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T17:30:16,653 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260 2024-12-02T17:30:16,656 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/zookeeper_0, clientPort=58907, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T17:30:16,657 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=58907 2024-12-02T17:30:16,657 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:30:16,658 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:30:16,666 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:30:16,666 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741825_1001 (size=7) 2024-12-02T17:30:16,667 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a with version=8 2024-12-02T17:30:16,667 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:38211/user/jenkins/test-data/a3543524-35fc-ec36-162b-f82ca9766da0/hbase-staging 2024-12-02T17:30:16,669 INFO [Time-limited test {}] client.ConnectionUtils(128): master/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:30:16,669 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:30:16,669 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:30:16,669 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:30:16,669 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:30:16,669 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:30:16,669 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-02T17:30:16,669 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:30:16,670 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:43883 2024-12-02T17:30:16,671 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:43883 connecting to ZooKeeper ensemble=127.0.0.1:58907 2024-12-02T17:30:16,676 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:438830x0, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:30:16,676 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:43883-0x1009c0a4d6e0000 connected 2024-12-02T17:30:16,690 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:30:16,691 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:30:16,693 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:30:16,693 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a, hbase.cluster.distributed=false 2024-12-02T17:30:16,694 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:30:16,695 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43883 2024-12-02T17:30:16,695 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43883 2024-12-02T17:30:16,695 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43883 2024-12-02T17:30:16,695 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43883 2024-12-02T17:30:16,695 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43883 2024-12-02T17:30:16,709 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/3b3375d1a925:0 server-side Connection retries=45 2024-12-02T17:30:16,709 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:30:16,709 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T17:30:16,709 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T17:30:16,709 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T17:30:16,709 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T17:30:16,709 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T17:30:16,710 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T17:30:16,710 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:44607 2024-12-02T17:30:16,711 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:44607 connecting to ZooKeeper ensemble=127.0.0.1:58907 2024-12-02T17:30:16,711 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:30:16,713 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:30:16,716 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:446070x0, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T17:30:16,717 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:446070x0, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:30:16,717 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:44607-0x1009c0a4d6e0001 connected 2024-12-02T17:30:16,717 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T17:30:16,717 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T17:30:16,718 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T17:30:16,719 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T17:30:16,719 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=44607 2024-12-02T17:30:16,719 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=44607 2024-12-02T17:30:16,719 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=44607 2024-12-02T17:30:16,720 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=44607 2024-12-02T17:30:16,720 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=44607 2024-12-02T17:30:16,730 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;3b3375d1a925:43883 2024-12-02T17:30:16,731 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/3b3375d1a925,43883,1733160616669 2024-12-02T17:30:16,732 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:30:16,732 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:30:16,733 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/3b3375d1a925,43883,1733160616669 2024-12-02T17:30:16,734 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T17:30:16,734 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,734 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,735 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T17:30:16,735 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/3b3375d1a925,43883,1733160616669 from backup master directory 2024-12-02T17:30:16,736 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/3b3375d1a925,43883,1733160616669 2024-12-02T17:30:16,736 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:30:16,736 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T17:30:16,736 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:30:16,736 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=3b3375d1a925,43883,1733160616669 2024-12-02T17:30:16,742 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/hbase.id] with ID: 22291efb-1a04-4710-9c8b-6e8bd880cf08 2024-12-02T17:30:16,742 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/.tmp/hbase.id 2024-12-02T17:30:16,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:30:16,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741826_1002 (size=42) 2024-12-02T17:30:16,748 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/.tmp/hbase.id]:[hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/hbase.id] 2024-12-02T17:30:16,758 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:30:16,758 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-02T17:30:16,759 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 1ms. 2024-12-02T17:30:16,761 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,761 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:30:16,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741827_1003 (size=196) 2024-12-02T17:30:16,767 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T17:30:16,767 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T17:30:16,768 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:30:16,773 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:30:16,773 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741828_1004 (size=1189) 2024-12-02T17:30:16,774 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store 2024-12-02T17:30:16,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:30:16,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741829_1005 (size=34) 2024-12-02T17:30:16,780 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:30:16,780 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:30:16,780 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:16,780 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:16,780 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:30:16,780 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:16,780 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:16,780 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160616780Disabling compacts and flushes for region at 1733160616780Disabling writes for close at 1733160616780Writing region close event to WAL at 1733160616780Closed at 1733160616780 2024-12-02T17:30:16,781 WARN [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/.initializing 2024-12-02T17:30:16,781 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/WALs/3b3375d1a925,43883,1733160616669 2024-12-02T17:30:16,783 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C43883%2C1733160616669, suffix=, logDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/WALs/3b3375d1a925,43883,1733160616669, archiveDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/oldWALs, maxLogs=10 2024-12-02T17:30:16,783 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C43883%2C1733160616669.1733160616783 2024-12-02T17:30:16,787 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/WALs/3b3375d1a925,43883,1733160616669/3b3375d1a925%2C43883%2C1733160616669.1733160616783 2024-12-02T17:30:16,787 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46875:46875),(127.0.0.1/127.0.0.1:34697:34697)] 2024-12-02T17:30:16,788 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:30:16,789 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:30:16,789 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,789 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,790 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,791 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T17:30:16,791 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,791 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:16,791 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,792 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T17:30:16,792 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,793 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:30:16,793 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,794 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T17:30:16,794 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,794 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:30:16,794 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,795 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T17:30:16,795 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,795 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T17:30:16,795 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,796 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,796 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,797 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,797 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,798 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T17:30:16,799 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T17:30:16,800 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:30:16,801 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=691529, jitterRate=-0.1206756979227066}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T17:30:16,801 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733160616789Initializing all the Stores at 1733160616789Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160616789Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160616790 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160616790Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160616790Cleaning up temporary data from old regions at 1733160616797 (+7 ms)Region opened successfully at 1733160616801 (+4 ms) 2024-12-02T17:30:16,801 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T17:30:16,804 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@19bc88c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:30:16,805 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-02T17:30:16,805 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T17:30:16,805 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T17:30:16,805 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T17:30:16,805 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T17:30:16,806 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-02T17:30:16,806 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T17:30:16,810 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T17:30:16,811 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T17:30:16,812 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-02T17:30:16,813 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T17:30:16,813 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T17:30:16,814 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-02T17:30:16,815 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T17:30:16,815 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T17:30:16,817 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-02T17:30:16,818 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T17:30:16,819 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T17:30:16,821 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T17:30:16,822 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T17:30:16,824 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:30:16,824 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T17:30:16,824 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,824 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,824 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=3b3375d1a925,43883,1733160616669, sessionid=0x1009c0a4d6e0000, setting cluster-up flag (Was=false) 2024-12-02T17:30:16,828 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,828 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,833 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T17:30:16,833 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,43883,1733160616669 2024-12-02T17:30:16,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:16,841 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T17:30:16,842 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=3b3375d1a925,43883,1733160616669 2024-12-02T17:30:16,843 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-02T17:30:16,844 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-02T17:30:16,844 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-02T17:30:16,844 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T17:30:16,845 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 3b3375d1a925,43883,1733160616669 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T17:30:16,845 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:30:16,845 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:30:16,846 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:30:16,846 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/3b3375d1a925:0, corePoolSize=5, maxPoolSize=5 2024-12-02T17:30:16,846 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/3b3375d1a925:0, corePoolSize=10, maxPoolSize=10 2024-12-02T17:30:16,846 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,846 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:30:16,846 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,846 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733160646846 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T17:30:16,847 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:30:16,847 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T17:30:16,847 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T17:30:16,848 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T17:30:16,848 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T17:30:16,848 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T17:30:16,848 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,848 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160616848,5,FailOnTimeoutGroup] 2024-12-02T17:30:16,848 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T17:30:16,848 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160616848,5,FailOnTimeoutGroup] 2024-12-02T17:30:16,848 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,848 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T17:30:16,848 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,849 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:30:16,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741831_1007 (size=1321) 2024-12-02T17:30:16,855 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-02T17:30:16,855 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a 2024-12-02T17:30:16,863 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:30:16,863 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741832_1008 (size=32) 2024-12-02T17:30:16,863 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:30:16,864 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:30:16,865 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:30:16,865 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,866 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:16,866 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:30:16,867 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:30:16,867 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,867 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:16,867 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:30:16,868 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:30:16,868 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,868 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:16,868 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:30:16,869 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:30:16,869 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:16,870 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:16,870 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:30:16,870 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740 2024-12-02T17:30:16,871 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740 2024-12-02T17:30:16,872 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:30:16,872 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:30:16,872 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:30:16,873 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:30:16,875 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T17:30:16,875 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=721597, jitterRate=-0.08244308829307556}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:30:16,875 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733160616863Initializing all the Stores at 1733160616864 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160616864Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160616864Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160616864Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160616864Cleaning up temporary data from old regions at 1733160616872 (+8 ms)Region opened successfully at 1733160616875 (+3 ms) 2024-12-02T17:30:16,876 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:30:16,876 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:30:16,876 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:30:16,876 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:30:16,876 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:30:16,876 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:30:16,876 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160616875Disabling compacts and flushes for region at 1733160616875Disabling writes for close at 1733160616876 (+1 ms)Writing region close event to WAL at 1733160616876Closed at 1733160616876 2024-12-02T17:30:16,877 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:30:16,877 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-02T17:30:16,877 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T17:30:16,878 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:30:16,879 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T17:30:16,921 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(746): ClusterId : 22291efb-1a04-4710-9c8b-6e8bd880cf08 2024-12-02T17:30:16,921 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T17:30:16,924 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T17:30:16,924 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T17:30:16,926 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T17:30:16,926 DEBUG [RS:0;3b3375d1a925:44607 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@65b5fe2a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=3b3375d1a925/172.17.0.3:0 2024-12-02T17:30:16,938 DEBUG [RS:0;3b3375d1a925:44607 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;3b3375d1a925:44607 2024-12-02T17:30:16,938 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-02T17:30:16,938 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-02T17:30:16,938 DEBUG [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-02T17:30:16,938 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(2659): reportForDuty to master=3b3375d1a925,43883,1733160616669 with port=44607, startcode=1733160616709 2024-12-02T17:30:16,938 DEBUG [RS:0;3b3375d1a925:44607 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T17:30:16,940 INFO [HMaster-EventLoopGroup-16-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:50409, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.7 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T17:30:16,941 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43883 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 3b3375d1a925,44607,1733160616709 2024-12-02T17:30:16,941 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43883 {}] master.ServerManager(517): Registering regionserver=3b3375d1a925,44607,1733160616709 2024-12-02T17:30:16,942 DEBUG [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a 2024-12-02T17:30:16,942 DEBUG [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:39979 2024-12-02T17:30:16,942 DEBUG [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-02T17:30:16,943 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:30:16,944 DEBUG [RS:0;3b3375d1a925:44607 {}] zookeeper.ZKUtil(111): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/3b3375d1a925,44607,1733160616709 2024-12-02T17:30:16,944 WARN [RS:0;3b3375d1a925:44607 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T17:30:16,944 INFO [RS:0;3b3375d1a925:44607 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:30:16,944 DEBUG [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/3b3375d1a925,44607,1733160616709 2024-12-02T17:30:16,944 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [3b3375d1a925,44607,1733160616709] 2024-12-02T17:30:16,947 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T17:30:16,948 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T17:30:16,948 INFO [RS:0;3b3375d1a925:44607 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T17:30:16,948 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,948 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-02T17:30:16,949 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-02T17:30:16,949 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/3b3375d1a925:0, corePoolSize=2, maxPoolSize=2 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/3b3375d1a925:0, corePoolSize=1, maxPoolSize=1 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:30:16,949 DEBUG [RS:0;3b3375d1a925:44607 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/3b3375d1a925:0, corePoolSize=3, maxPoolSize=3 2024-12-02T17:30:16,950 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,950 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,950 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,950 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,950 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,950 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,44607,1733160616709-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:30:16,963 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T17:30:16,963 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,44607,1733160616709-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,963 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,964 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.Replication(171): 3b3375d1a925,44607,1733160616709 started 2024-12-02T17:30:16,977 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:16,977 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1482): Serving as 3b3375d1a925,44607,1733160616709, RpcServer on 3b3375d1a925/172.17.0.3:44607, sessionid=0x1009c0a4d6e0001 2024-12-02T17:30:16,977 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T17:30:16,977 DEBUG [RS:0;3b3375d1a925:44607 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 3b3375d1a925,44607,1733160616709 2024-12-02T17:30:16,977 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,44607,1733160616709' 2024-12-02T17:30:16,977 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T17:30:16,977 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T17:30:16,978 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T17:30:16,978 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T17:30:16,978 DEBUG [RS:0;3b3375d1a925:44607 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 3b3375d1a925,44607,1733160616709 2024-12-02T17:30:16,978 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '3b3375d1a925,44607,1733160616709' 2024-12-02T17:30:16,978 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T17:30:16,978 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T17:30:16,978 DEBUG [RS:0;3b3375d1a925:44607 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T17:30:16,978 INFO [RS:0;3b3375d1a925:44607 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T17:30:16,978 INFO [RS:0;3b3375d1a925:44607 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T17:30:17,029 WARN [3b3375d1a925:43883 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-02T17:30:17,080 INFO [RS:0;3b3375d1a925:44607 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C44607%2C1733160616709, suffix=, logDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/3b3375d1a925,44607,1733160616709, archiveDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/oldWALs, maxLogs=32 2024-12-02T17:30:17,081 INFO [RS:0;3b3375d1a925:44607 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C44607%2C1733160616709.1733160617081 2024-12-02T17:30:17,102 INFO [RS:0;3b3375d1a925:44607 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/3b3375d1a925,44607,1733160616709/3b3375d1a925%2C44607%2C1733160616709.1733160617081 2024-12-02T17:30:17,104 DEBUG [RS:0;3b3375d1a925:44607 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34697:34697),(127.0.0.1/127.0.0.1:46875:46875)] 2024-12-02T17:30:17,279 DEBUG [3b3375d1a925:43883 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T17:30:17,280 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=3b3375d1a925,44607,1733160616709 2024-12-02T17:30:17,281 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,44607,1733160616709, state=OPENING 2024-12-02T17:30:17,283 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T17:30:17,284 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:17,284 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:17,285 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T17:30:17,285 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:30:17,285 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:30:17,285 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,44607,1733160616709}] 2024-12-02T17:30:17,438 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T17:30:17,440 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-17-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:60505, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T17:30:17,443 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-02T17:30:17,443 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:30:17,445 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=3b3375d1a925%2C44607%2C1733160616709.meta, suffix=.meta, logDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/3b3375d1a925,44607,1733160616709, archiveDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/oldWALs, maxLogs=32 2024-12-02T17:30:17,445 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 3b3375d1a925%2C44607%2C1733160616709.meta.1733160617445.meta 2024-12-02T17:30:17,451 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/3b3375d1a925,44607,1733160616709/3b3375d1a925%2C44607%2C1733160616709.meta.1733160617445.meta 2024-12-02T17:30:17,457 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46875:46875),(127.0.0.1/127.0.0.1:34697:34697)] 2024-12-02T17:30:17,462 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T17:30:17,462 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T17:30:17,462 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T17:30:17,462 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T17:30:17,462 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T17:30:17,462 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T17:30:17,462 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-02T17:30:17,462 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-02T17:30:17,464 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T17:30:17,464 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T17:30:17,464 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:17,465 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:17,465 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-02T17:30:17,465 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-02T17:30:17,465 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:17,466 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:17,466 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T17:30:17,466 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T17:30:17,466 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:17,467 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:17,467 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T17:30:17,467 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T17:30:17,467 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T17:30:17,468 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T17:30:17,468 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-02T17:30:17,468 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740 2024-12-02T17:30:17,469 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740 2024-12-02T17:30:17,470 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-02T17:30:17,470 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-02T17:30:17,471 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T17:30:17,472 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-02T17:30:17,472 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=801387, jitterRate=0.019016817212104797}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T17:30:17,472 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-02T17:30:17,473 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733160617463Writing region info on filesystem at 1733160617463Initializing all the Stores at 1733160617463Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160617463Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160617463Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733160617463Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733160617463Cleaning up temporary data from old regions at 1733160617470 (+7 ms)Running coprocessor post-open hooks at 1733160617472 (+2 ms)Region opened successfully at 1733160617473 (+1 ms) 2024-12-02T17:30:17,474 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733160617437 2024-12-02T17:30:17,476 DEBUG [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T17:30:17,476 INFO [RS_OPEN_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-02T17:30:17,477 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=3b3375d1a925,44607,1733160616709 2024-12-02T17:30:17,477 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 3b3375d1a925,44607,1733160616709, state=OPEN 2024-12-02T17:30:17,481 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:30:17,481 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T17:30:17,482 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=3b3375d1a925,44607,1733160616709 2024-12-02T17:30:17,482 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:30:17,482 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T17:30:17,484 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T17:30:17,484 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=3b3375d1a925,44607,1733160616709 in 197 msec 2024-12-02T17:30:17,486 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T17:30:17,486 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 607 msec 2024-12-02T17:30:17,486 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-02T17:30:17,486 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-02T17:30:17,488 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:30:17,488 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,44607,1733160616709, seqNum=-1] 2024-12-02T17:30:17,488 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:30:17,489 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-17-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51561, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:30:17,493 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 649 msec 2024-12-02T17:30:17,493 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733160617493, completionTime=-1 2024-12-02T17:30:17,493 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T17:30:17,493 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-02T17:30:17,495 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=1 2024-12-02T17:30:17,495 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733160677495 2024-12-02T17:30:17,495 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733160737495 2024-12-02T17:30:17,495 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 1 msec 2024-12-02T17:30:17,496 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,43883,1733160616669-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:17,496 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,43883,1733160616669-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:17,496 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,43883,1733160616669-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:17,496 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-3b3375d1a925:43883, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:17,496 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:17,496 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:17,497 DEBUG [master/3b3375d1a925:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-02T17:30:17,499 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.763sec 2024-12-02T17:30:17,499 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T17:30:17,499 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T17:30:17,500 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T17:30:17,500 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T17:30:17,500 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T17:30:17,500 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,43883,1733160616669-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T17:30:17,500 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,43883,1733160616669-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T17:30:17,502 DEBUG [master/3b3375d1a925:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-02T17:30:17,502 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T17:30:17,502 INFO [master/3b3375d1a925:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=3b3375d1a925,43883,1733160616669-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T17:30:17,522 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@32b4a64d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:30:17,522 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 3b3375d1a925,43883,-1 for getting cluster id 2024-12-02T17:30:17,522 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-02T17:30:17,523 DEBUG [HMaster-EventLoopGroup-16-3 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '22291efb-1a04-4710-9c8b-6e8bd880cf08' 2024-12-02T17:30:17,523 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-02T17:30:17,524 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "22291efb-1a04-4710-9c8b-6e8bd880cf08" 2024-12-02T17:30:17,524 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@144b859b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:30:17,524 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [3b3375d1a925,43883,-1] 2024-12-02T17:30:17,524 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-02T17:30:17,524 DEBUG [RPCClient-NioEventLoopGroup-4-8 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:17,525 INFO [HMaster-EventLoopGroup-16-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54364, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-02T17:30:17,526 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@616bab4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T17:30:17,526 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-02T17:30:17,527 DEBUG [RPCClient-NioEventLoopGroup-4-9 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=3b3375d1a925,44607,1733160616709, seqNum=-1] 2024-12-02T17:30:17,527 DEBUG [RPCClient-NioEventLoopGroup-4-9 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T17:30:17,528 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-17-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56606, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T17:30:17,529 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=3b3375d1a925,43883,1733160616669 2024-12-02T17:30:17,529 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T17:30:17,532 INFO [Time-limited test {}] master.MasterRpcServices(567): Client=null/null set balanceSwitch=false 2024-12-02T17:30:17,532 INFO [Time-limited test {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T17:30:17,534 INFO [Time-limited test {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=test.com%2C8080%2C1, suffix=, logDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/test.com,8080,1, archiveDir=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/oldWALs, maxLogs=32 2024-12-02T17:30:17,534 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor test.com%2C8080%2C1.1733160617534 2024-12-02T17:30:17,539 INFO [Time-limited test {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/test.com,8080,1/test.com%2C8080%2C1.1733160617534 2024-12-02T17:30:17,545 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34697:34697),(127.0.0.1/127.0.0.1:46875:46875)] 2024-12-02T17:30:17,548 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor test.com%2C8080%2C1.1733160617548 2024-12-02T17:30:17,552 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,553 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,553 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,553 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,553 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,553 INFO [Time-limited test {}] wal.AbstractFSWAL(987): Rolled WAL /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/test.com,8080,1/test.com%2C8080%2C1.1733160617534 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/test.com,8080,1/test.com%2C8080%2C1.1733160617548 2024-12-02T17:30:17,554 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34697:34697),(127.0.0.1/127.0.0.1:46875:46875)] 2024-12-02T17:30:17,554 DEBUG [Time-limited test {}] wal.AbstractFSWAL(879): hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/test.com,8080,1/test.com%2C8080%2C1.1733160617534 is not closed yet, will try archiving it next time 2024-12-02T17:30:17,554 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,554 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741835_1011 (size=93) 2024-12-02T17:30:17,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741835_1011 (size=93) 2024-12-02T17:30:17,555 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,555 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,555 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,556 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(968): Archiving hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/WALs/test.com,8080,1/test.com%2C8080%2C1.1733160617534 to hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/oldWALs/test.com%2C8080%2C1.1733160617534 2024-12-02T17:30:17,557 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741836_1012 (size=93) 2024-12-02T17:30:17,557 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741836_1012 (size=93) 2024-12-02T17:30:17,559 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/oldWALs 2024-12-02T17:30:17,559 INFO [Time-limited test {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog test.com%2C8080%2C1:(num 1733160617548) 2024-12-02T17:30:17,559 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-02T17:30:17,559 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:30:17,559 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:30:17,559 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:17,559 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:17,559 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-02T17:30:17,559 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T17:30:17,560 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=190723906, stopped=false 2024-12-02T17:30:17,560 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=3b3375d1a925,43883,1733160616669 2024-12-02T17:30:17,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:30:17,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T17:30:17,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:17,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:17,561 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:30:17,561 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-02T17:30:17,561 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.AbstractTestLogRolling.tearDown(AbstractTestLogRolling.java:163) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:30:17,561 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:17,562 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '3b3375d1a925,44607,1733160616709' ***** 2024-12-02T17:30:17,562 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-02T17:30:17,562 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T17:30:17,562 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T17:30:17,562 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(959): stopping server 3b3375d1a925,44607,1733160616709 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;3b3375d1a925:44607. 2024-12-02T17:30:17,562 DEBUG [RS:0;3b3375d1a925:44607 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-02T17:30:17,562 DEBUG [RS:0;3b3375d1a925:44607 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T17:30:17,562 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T17:30:17,563 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-02T17:30:17,563 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-02T17:30:17,563 DEBUG [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-02T17:30:17,563 DEBUG [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-02T17:30:17,563 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-02T17:30:17,563 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-02T17:30:17,563 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-02T17:30:17,563 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T17:30:17,563 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T17:30:17,563 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=74 B heapSize=1.22 KB 2024-12-02T17:30:17,579 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740/.tmp/ns/ebd247db12d54faaa616a43bbf3bcda0 is 43, key is default/ns:d/1733160617490/Put/seqid=0 2024-12-02T17:30:17,583 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741837_1013 (size=5153) 2024-12-02T17:30:17,584 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741837_1013 (size=5153) 2024-12-02T17:30:17,584 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740/.tmp/ns/ebd247db12d54faaa616a43bbf3bcda0 2024-12-02T17:30:17,590 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740/.tmp/ns/ebd247db12d54faaa616a43bbf3bcda0 as hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740/ns/ebd247db12d54faaa616a43bbf3bcda0 2024-12-02T17:30:17,594 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740/ns/ebd247db12d54faaa616a43bbf3bcda0, entries=2, sequenceid=6, filesize=5.0 K 2024-12-02T17:30:17,595 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~74 B/74, heapSize ~464 B/464, currentSize=0 B/0 for 1588230740 in 32ms, sequenceid=6, compaction requested=false 2024-12-02T17:30:17,595 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T17:30:17,599 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/data/hbase/meta/1588230740/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-02T17:30:17,599 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:30:17,600 INFO [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-02T17:30:17,600 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733160617563Running coprocessor pre-close hooks at 1733160617563Disabling compacts and flushes for region at 1733160617563Disabling writes for close at 1733160617563Obtaining lock to block concurrent updates at 1733160617563Preparing flush snapshotting stores in 1588230740 at 1733160617563Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=74, getHeapSize=1184, getOffHeapSize=0, getCellsCount=2 at 1733160617563Flushing stores of hbase:meta,,1.1588230740 at 1733160617564 (+1 ms)Flushing 1588230740/ns: creating writer at 1733160617564Flushing 1588230740/ns: appending metadata at 1733160617579 (+15 ms)Flushing 1588230740/ns: closing flushed file at 1733160617579Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4b6f170b: reopening flushed file at 1733160617589 (+10 ms)Finished flush of dataSize ~74 B/74, heapSize ~464 B/464, currentSize=0 B/0 for 1588230740 in 32ms, sequenceid=6, compaction requested=false at 1733160617595 (+6 ms)Writing region close event to WAL at 1733160617596 (+1 ms)Running coprocessor post-close hooks at 1733160617599 (+3 ms)Closed at 1733160617599 2024-12-02T17:30:17,600 DEBUG [RS_CLOSE_META-regionserver/3b3375d1a925:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T17:30:17,763 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(976): stopping server 3b3375d1a925,44607,1733160616709; all regions closed. 2024-12-02T17:30:17,764 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,764 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,764 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,764 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,764 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741834_1010 (size=1152) 2024-12-02T17:30:17,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741834_1010 (size=1152) 2024-12-02T17:30:17,768 DEBUG [RS:0;3b3375d1a925:44607 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/oldWALs 2024-12-02T17:30:17,768 INFO [RS:0;3b3375d1a925:44607 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C44607%2C1733160616709.meta:.meta(num 1733160617445) 2024-12-02T17:30:17,769 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,769 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,769 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,769 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,769 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,770 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741833_1009 (size=93) 2024-12-02T17:30:17,771 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741833_1009 (size=93) 2024-12-02T17:30:17,772 DEBUG [RS:0;3b3375d1a925:44607 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/oldWALs 2024-12-02T17:30:17,772 INFO [RS:0;3b3375d1a925:44607 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 3b3375d1a925%2C44607%2C1733160616709:(num 1733160617081) 2024-12-02T17:30:17,772 DEBUG [RS:0;3b3375d1a925:44607 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T17:30:17,772 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T17:30:17,772 INFO [RS:0;3b3375d1a925:44607 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:30:17,773 INFO [RS:0;3b3375d1a925:44607 {}] hbase.ChoreService(370): Chore service for: regionserver/3b3375d1a925:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-02T17:30:17,773 INFO [RS:0;3b3375d1a925:44607 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:30:17,773 INFO [regionserver/3b3375d1a925:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:30:17,773 INFO [RS:0;3b3375d1a925:44607 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:44607 2024-12-02T17:30:17,775 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/3b3375d1a925,44607,1733160616709 2024-12-02T17:30:17,775 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T17:30:17,775 INFO [RS:0;3b3375d1a925:44607 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:30:17,776 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [3b3375d1a925,44607,1733160616709] 2024-12-02T17:30:17,778 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/3b3375d1a925,44607,1733160616709 already deleted, retry=false 2024-12-02T17:30:17,778 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 3b3375d1a925,44607,1733160616709 expired; onlineServers=0 2024-12-02T17:30:17,778 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '3b3375d1a925,43883,1733160616669' ***** 2024-12-02T17:30:17,778 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T17:30:17,778 INFO [M:0;3b3375d1a925:43883 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-02T17:30:17,778 INFO [M:0;3b3375d1a925:43883 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-02T17:30:17,778 DEBUG [M:0;3b3375d1a925:43883 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T17:30:17,778 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T17:30:17,778 DEBUG [M:0;3b3375d1a925:43883 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T17:30:17,778 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160616848 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.large.0-1733160616848,5,FailOnTimeoutGroup] 2024-12-02T17:30:17,778 DEBUG [master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160616848 {}] cleaner.HFileCleaner(306): Exit Thread[master/3b3375d1a925:0:becomeActiveMaster-HFileCleaner.small.0-1733160616848,5,FailOnTimeoutGroup] 2024-12-02T17:30:17,778 INFO [M:0;3b3375d1a925:43883 {}] hbase.ChoreService(370): Chore service for: master/3b3375d1a925:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-02T17:30:17,778 INFO [M:0;3b3375d1a925:43883 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-02T17:30:17,778 DEBUG [M:0;3b3375d1a925:43883 {}] master.HMaster(1795): Stopping service threads 2024-12-02T17:30:17,779 INFO [M:0;3b3375d1a925:43883 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T17:30:17,779 INFO [M:0;3b3375d1a925:43883 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-02T17:30:17,779 INFO [M:0;3b3375d1a925:43883 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T17:30:17,779 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T17:30:17,781 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T17:30:17,781 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T17:30:17,781 DEBUG [M:0;3b3375d1a925:43883 {}] zookeeper.ZKUtil(347): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T17:30:17,781 WARN [M:0;3b3375d1a925:43883 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T17:30:17,781 INFO [M:0;3b3375d1a925:43883 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/.lastflushedseqids 2024-12-02T17:30:17,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741838_1014 (size=99) 2024-12-02T17:30:17,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741838_1014 (size=99) 2024-12-02T17:30:17,786 INFO [M:0;3b3375d1a925:43883 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-02T17:30:17,786 INFO [M:0;3b3375d1a925:43883 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T17:30:17,787 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T17:30:17,787 INFO [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:17,787 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:17,787 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T17:30:17,787 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:17,787 INFO [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=7.67 KB heapSize=11.34 KB 2024-12-02T17:30:17,802 DEBUG [M:0;3b3375d1a925:43883 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/b2b570c5028e496f8f84700f4ed09bbc is 82, key is hbase:meta,,1/info:regioninfo/1733160617476/Put/seqid=0 2024-12-02T17:30:17,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741839_1015 (size=5672) 2024-12-02T17:30:17,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741839_1015 (size=5672) 2024-12-02T17:30:17,806 INFO [M:0;3b3375d1a925:43883 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=29 (bloomFilter=true), to=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/b2b570c5028e496f8f84700f4ed09bbc 2024-12-02T17:30:17,823 DEBUG [M:0;3b3375d1a925:43883 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1b3c3595f1744a59a69183802dc3253c is 240, key is \x00\x00\x00\x00\x00\x00\x00\x01/proc:d/1733160617493/Put/seqid=0 2024-12-02T17:30:17,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741840_1016 (size=5275) 2024-12-02T17:30:17,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741840_1016 (size=5275) 2024-12-02T17:30:17,828 INFO [M:0;3b3375d1a925:43883 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.06 KB at sequenceid=29 (bloomFilter=true), to=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1b3c3595f1744a59a69183802dc3253c 2024-12-02T17:30:17,845 DEBUG [M:0;3b3375d1a925:43883 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6542ce2652ab45e98dd2a4bfac420897 is 69, key is 3b3375d1a925,44607,1733160616709/rs:state/1733160616941/Put/seqid=0 2024-12-02T17:30:17,849 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741841_1017 (size=5156) 2024-12-02T17:30:17,850 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741841_1017 (size=5156) 2024-12-02T17:30:17,850 INFO [M:0;3b3375d1a925:43883 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=29 (bloomFilter=true), to=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6542ce2652ab45e98dd2a4bfac420897 2024-12-02T17:30:17,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T17:30:17,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T17:30:17,867 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRolling 2024-12-02T17:30:17,867 DEBUG [M:0;3b3375d1a925:43883 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/eaee0827f4ce425f9936c73cbcc36a71 is 52, key is load_balancer_on/state:d/1733160617531/Put/seqid=0 2024-12-02T17:30:17,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741842_1018 (size=5056) 2024-12-02T17:30:17,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741842_1018 (size=5056) 2024-12-02T17:30:17,872 INFO [M:0;3b3375d1a925:43883 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=29 (bloomFilter=true), to=hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/eaee0827f4ce425f9936c73cbcc36a71 2024-12-02T17:30:17,876 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/b2b570c5028e496f8f84700f4ed09bbc as hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/b2b570c5028e496f8f84700f4ed09bbc 2024-12-02T17:30:17,876 INFO [RS:0;3b3375d1a925:44607 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:30:17,876 INFO [RS:0;3b3375d1a925:44607 {}] regionserver.HRegionServer(1031): Exiting; stopping=3b3375d1a925,44607,1733160616709; zookeeper connection closed. 2024-12-02T17:30:17,877 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:30:17,877 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44607-0x1009c0a4d6e0001, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:30:17,877 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7d245a11 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7d245a11 2024-12-02T17:30:17,877 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T17:30:17,880 INFO [M:0;3b3375d1a925:43883 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/b2b570c5028e496f8f84700f4ed09bbc, entries=8, sequenceid=29, filesize=5.5 K 2024-12-02T17:30:17,881 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1b3c3595f1744a59a69183802dc3253c as hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/1b3c3595f1744a59a69183802dc3253c 2024-12-02T17:30:17,884 INFO [M:0;3b3375d1a925:43883 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/1b3c3595f1744a59a69183802dc3253c, entries=3, sequenceid=29, filesize=5.2 K 2024-12-02T17:30:17,885 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6542ce2652ab45e98dd2a4bfac420897 as hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/6542ce2652ab45e98dd2a4bfac420897 2024-12-02T17:30:17,889 INFO [M:0;3b3375d1a925:43883 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/6542ce2652ab45e98dd2a4bfac420897, entries=1, sequenceid=29, filesize=5.0 K 2024-12-02T17:30:17,889 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/eaee0827f4ce425f9936c73cbcc36a71 as hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/eaee0827f4ce425f9936c73cbcc36a71 2024-12-02T17:30:17,893 INFO [M:0;3b3375d1a925:43883 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:39979/user/jenkins/test-data/f7e2b370-d7e1-2634-2d7f-c773b234ad1a/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/eaee0827f4ce425f9936c73cbcc36a71, entries=1, sequenceid=29, filesize=4.9 K 2024-12-02T17:30:17,894 INFO [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(3140): Finished flush of dataSize ~7.67 KB/7850, heapSize ~11.27 KB/11544, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 107ms, sequenceid=29, compaction requested=false 2024-12-02T17:30:17,895 INFO [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T17:30:17,895 DEBUG [M:0;3b3375d1a925:43883 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733160617787Disabling compacts and flushes for region at 1733160617787Disabling writes for close at 1733160617787Obtaining lock to block concurrent updates at 1733160617787Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733160617787Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=7850, getHeapSize=11544, getOffHeapSize=0, getCellsCount=36 at 1733160617787Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733160617788 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733160617788Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733160617802 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733160617802Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733160617810 (+8 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733160617823 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733160617823Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733160617832 (+9 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733160617845 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733160617845Flushing 1595e783b53d99cd5eef43b6debb2682/state: creating writer at 1733160617853 (+8 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: appending metadata at 1733160617867 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/state: closing flushed file at 1733160617867Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@192d288a: reopening flushed file at 1733160617875 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@2f56d2b1: reopening flushed file at 1733160617880 (+5 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@17effa89: reopening flushed file at 1733160617884 (+4 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@57150d3f: reopening flushed file at 1733160617889 (+5 ms)Finished flush of dataSize ~7.67 KB/7850, heapSize ~11.27 KB/11544, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 107ms, sequenceid=29, compaction requested=false at 1733160617894 (+5 ms)Writing region close event to WAL at 1733160617895 (+1 ms)Closed at 1733160617895 2024-12-02T17:30:17,895 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,896 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,896 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,896 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,896 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-02T17:30:17,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33331 is added to blk_1073741830_1006 (size=10311) 2024-12-02T17:30:17,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41719 is added to blk_1073741830_1006 (size=10311) 2024-12-02T17:30:17,898 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-02T17:30:17,898 INFO [M:0;3b3375d1a925:43883 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-02T17:30:17,898 INFO [M:0;3b3375d1a925:43883 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:43883 2024-12-02T17:30:17,899 INFO [M:0;3b3375d1a925:43883 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-02T17:30:18,001 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:30:18,001 INFO [M:0;3b3375d1a925:43883 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-02T17:30:18,001 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43883-0x1009c0a4d6e0000, quorum=127.0.0.1:58907, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T17:30:18,003 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@a8a2fb4{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:30:18,003 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1f41372a{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:30:18,004 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:30:18,004 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@e9b8f9f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:30:18,004 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2b13a29{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/hadoop.log.dir/,STOPPED} 2024-12-02T17:30:18,005 WARN [BP-1495924061-172.17.0.3-1733160616036 heartbeating to localhost/127.0.0.1:39979 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:30:18,005 WARN [BP-1495924061-172.17.0.3-1733160616036 heartbeating to localhost/127.0.0.1:39979 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1495924061-172.17.0.3-1733160616036 (Datanode Uuid 6ec327de-6496-47da-aa4f-b4f73202aa4f) service to localhost/127.0.0.1:39979 2024-12-02T17:30:18,005 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:30:18,005 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:30:18,006 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/data/data3/current/BP-1495924061-172.17.0.3-1733160616036 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:30:18,006 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/data/data4/current/BP-1495924061-172.17.0.3-1733160616036 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:30:18,006 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:30:18,008 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@69a2ae1b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T17:30:18,008 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@64e1b9c7{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:30:18,008 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:30:18,008 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@40f3733a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:30:18,008 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@18478920{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/hadoop.log.dir/,STOPPED} 2024-12-02T17:30:18,010 WARN [BP-1495924061-172.17.0.3-1733160616036 heartbeating to localhost/127.0.0.1:39979 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T17:30:18,010 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T17:30:18,010 WARN [BP-1495924061-172.17.0.3-1733160616036 heartbeating to localhost/127.0.0.1:39979 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1495924061-172.17.0.3-1733160616036 (Datanode Uuid 7effab17-28f7-4f2b-bef3-9d3e9a5fae23) service to localhost/127.0.0.1:39979 2024-12-02T17:30:18,010 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T17:30:18,010 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/data/data1/current/BP-1495924061-172.17.0.3-1733160616036 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:30:18,010 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/cluster_4f3f7e8e-01bc-1f17-c129-ab1aa8261380/data/data2/current/BP-1495924061-172.17.0.3-1733160616036 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T17:30:18,010 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T17:30:18,016 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7b063aa1{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T17:30:18,016 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@778227aa{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T17:30:18,016 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T17:30:18,016 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4d118eec{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T17:30:18,017 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3073e97e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/0a023e16-09ff-b414-02d3-b38836200260/hadoop.log.dir/,STOPPED} 2024-12-02T17:30:18,024 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-02T17:30:18,038 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-02T17:30:18,048 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnNothingWritten Thread=266 (was 224) Potentially hanging thread: nioEventLoopGroup-43-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:39979 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-42-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-16-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-17-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.7@localhost:39979 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-42-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-42-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-45-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-45-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-43-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:39979 from jenkins.hfs.7 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:39979 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-44-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: globalEventExecutor-1-21 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//io.netty.util.concurrent.GlobalEventExecutor.takeTask(GlobalEventExecutor.java:113) app//io.netty.util.concurrent.GlobalEventExecutor$TaskRunner.run(GlobalEventExecutor.java:259) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (75657370) connection to localhost/127.0.0.1:39979 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:39979 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-44-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-44-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:39979 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-43-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-17-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: MiniHBaseClusterRegionServer-EventLoopGroup-17-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HMaster-EventLoopGroup-16-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-45-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:39979 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: HMaster-EventLoopGroup-16-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=536 (was 509) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=137 (was 137), ProcessCount=11 (was 11), AvailableMemoryMB=3092 (was 3105)